maxxxzdn commited on
Commit
127cde1
verified
1 Parent(s): 78795d9

README: move Citation to the very end

Browse files
Files changed (1) hide show
  1. README.md +14 -14
README.md CHANGED
@@ -216,20 +216,6 @@ gs://weatherbench2/datasets/hres_t0/2016-2022-6h-240x121_equiangular_with_poles_
216
 
217
  Mosaic operates at 1.5掳 (~166 km), which cannot resolve mesoscale phenomena such as tropical-cyclone inner-core structure or individual severe thunderstorms. The block-sparse attention is designed to scale linearly with sequence length, so finer grids (e.g. 0.25掳, ~700k tokens) are a natural next step but are not part of this release.
218
 
219
- ## Citation
220
-
221
- If you use Mosaic, please cite:
222
-
223
- ```bibtex
224
- @inproceedings{zhdanov2026mosaic,
225
- title = {(Sparse) Attention to the Details: Preserving Spectral Fidelity in ML-based Weather Forecasting Models},
226
- author = {Zhdanov, Maksim and Lucic, Ana and Welling, Max and van de Meent, Jan-Willem},
227
- booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
228
- year = {2026},
229
- url = {https://arxiv.org/abs/2604.16429}
230
- }
231
- ```
232
-
233
  ## License
234
 
235
  Released under [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/). Free for non-commercial research and educational use with attribution; commercial use requires a separate license. Underlying training data (ERA5, HRES) is subject to its own licensing terms set by ECMWF.
@@ -245,3 +231,17 @@ MZ acknowledges support from Microsoft Research AI4Science. JWvdM acknowledges s
245
  | License | [`cc-by-nc-4.0`](https://creativecommons.org/licenses/by-nc/4.0/) |
246
  | Library | `pytorch` |
247
  | Tags | `weather` 路 `weather-forecasting` 路 `climate` 路 `atmospheric-science` 路 `sparse-attention` 路 `transformer` 路 `probabilistic-forecasting` |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
216
 
217
  Mosaic operates at 1.5掳 (~166 km), which cannot resolve mesoscale phenomena such as tropical-cyclone inner-core structure or individual severe thunderstorms. The block-sparse attention is designed to scale linearly with sequence length, so finer grids (e.g. 0.25掳, ~700k tokens) are a natural next step but are not part of this release.
218
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
219
  ## License
220
 
221
  Released under [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/). Free for non-commercial research and educational use with attribution; commercial use requires a separate license. Underlying training data (ERA5, HRES) is subject to its own licensing terms set by ECMWF.
 
231
  | License | [`cc-by-nc-4.0`](https://creativecommons.org/licenses/by-nc/4.0/) |
232
  | Library | `pytorch` |
233
  | Tags | `weather` 路 `weather-forecasting` 路 `climate` 路 `atmospheric-science` 路 `sparse-attention` 路 `transformer` 路 `probabilistic-forecasting` |
234
+
235
+ ## Citation
236
+
237
+ If you use Mosaic, please cite:
238
+
239
+ ```bibtex
240
+ @inproceedings{zhdanov2026mosaic,
241
+ title = {(Sparse) Attention to the Details: Preserving Spectral Fidelity in ML-based Weather Forecasting Models},
242
+ author = {Zhdanov, Maksim and Lucic, Ana and Welling, Max and van de Meent, Jan-Willem},
243
+ booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
244
+ year = {2026},
245
+ url = {https://arxiv.org/abs/2604.16429}
246
+ }
247
+ ```