README: move Citation to the very end
Browse files
README.md
CHANGED
|
@@ -216,20 +216,6 @@ gs://weatherbench2/datasets/hres_t0/2016-2022-6h-240x121_equiangular_with_poles_
|
|
| 216 |
|
| 217 |
Mosaic operates at 1.5掳 (~166 km), which cannot resolve mesoscale phenomena such as tropical-cyclone inner-core structure or individual severe thunderstorms. The block-sparse attention is designed to scale linearly with sequence length, so finer grids (e.g. 0.25掳, ~700k tokens) are a natural next step but are not part of this release.
|
| 218 |
|
| 219 |
-
## Citation
|
| 220 |
-
|
| 221 |
-
If you use Mosaic, please cite:
|
| 222 |
-
|
| 223 |
-
```bibtex
|
| 224 |
-
@inproceedings{zhdanov2026mosaic,
|
| 225 |
-
title = {(Sparse) Attention to the Details: Preserving Spectral Fidelity in ML-based Weather Forecasting Models},
|
| 226 |
-
author = {Zhdanov, Maksim and Lucic, Ana and Welling, Max and van de Meent, Jan-Willem},
|
| 227 |
-
booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
|
| 228 |
-
year = {2026},
|
| 229 |
-
url = {https://arxiv.org/abs/2604.16429}
|
| 230 |
-
}
|
| 231 |
-
```
|
| 232 |
-
|
| 233 |
## License
|
| 234 |
|
| 235 |
Released under [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/). Free for non-commercial research and educational use with attribution; commercial use requires a separate license. Underlying training data (ERA5, HRES) is subject to its own licensing terms set by ECMWF.
|
|
@@ -245,3 +231,17 @@ MZ acknowledges support from Microsoft Research AI4Science. JWvdM acknowledges s
|
|
| 245 |
| License | [`cc-by-nc-4.0`](https://creativecommons.org/licenses/by-nc/4.0/) |
|
| 246 |
| Library | `pytorch` |
|
| 247 |
| Tags | `weather` 路 `weather-forecasting` 路 `climate` 路 `atmospheric-science` 路 `sparse-attention` 路 `transformer` 路 `probabilistic-forecasting` |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 216 |
|
| 217 |
Mosaic operates at 1.5掳 (~166 km), which cannot resolve mesoscale phenomena such as tropical-cyclone inner-core structure or individual severe thunderstorms. The block-sparse attention is designed to scale linearly with sequence length, so finer grids (e.g. 0.25掳, ~700k tokens) are a natural next step but are not part of this release.
|
| 218 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 219 |
## License
|
| 220 |
|
| 221 |
Released under [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/). Free for non-commercial research and educational use with attribution; commercial use requires a separate license. Underlying training data (ERA5, HRES) is subject to its own licensing terms set by ECMWF.
|
|
|
|
| 231 |
| License | [`cc-by-nc-4.0`](https://creativecommons.org/licenses/by-nc/4.0/) |
|
| 232 |
| Library | `pytorch` |
|
| 233 |
| Tags | `weather` 路 `weather-forecasting` 路 `climate` 路 `atmospheric-science` 路 `sparse-attention` 路 `transformer` 路 `probabilistic-forecasting` |
|
| 234 |
+
|
| 235 |
+
## Citation
|
| 236 |
+
|
| 237 |
+
If you use Mosaic, please cite:
|
| 238 |
+
|
| 239 |
+
```bibtex
|
| 240 |
+
@inproceedings{zhdanov2026mosaic,
|
| 241 |
+
title = {(Sparse) Attention to the Details: Preserving Spectral Fidelity in ML-based Weather Forecasting Models},
|
| 242 |
+
author = {Zhdanov, Maksim and Lucic, Ana and Welling, Max and van de Meent, Jan-Willem},
|
| 243 |
+
booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
|
| 244 |
+
year = {2026},
|
| 245 |
+
url = {https://arxiv.org/abs/2604.16429}
|
| 246 |
+
}
|
| 247 |
+
```
|