Poster (After Call) FZJ-2025-00343

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
Downscaling with the foundation model AtmoRep

 ;  ;  ;

2024

European Geosciences Union General Assembly 2024, EGU 2024, ViennaVienna, Austria, 14 Apr 2024 - 19 Apr 20242024-04-142024-04-19 [10.5194/egusphere-egu24-18331]

This record in other databases:

Please use a persistent id in citations: doi:  doi:

Abstract: In recent years, deep neural networks (DNN) to enhance the resolution of meteorological data, known as statistical downscaling, have surpassed classical statistical methods that have been developed previously with respect to several validation metrics. The prevailing approach for DNN downscaling is to train deep learning models in an end-to-end manner. However, foundation models trained on very large datasets in a self-supervised way have proven to provide new SOTA results for various applications in natural language processing and computer vision. To investigate the benefit of foundation models in Earth Science applications, we deploy the large-scale representation model for atmospheric dynamics AtmoRep (Lessig et al., 2023) for statistical downscaling of the 2m temperature over Central Europe. AtmoRep has been trained on almost 40 years of ERA5 data from 1979 to 2017 and has shown promising skill in several intrinsic and downstream applications. By extending AtmoRep’s encoder-decoder with a tail network for downscaling, we super-resolve the coarse-grained 2 m temperature field from ERA5-data (Δx = 25 km) to attain the high spatial resolution (Δx = 6 km) of the COSMO REA6 dataset. Different coupling approaches between the core and tail network (e.g. with and without fine-tuning the core model) are tested and analyzed in terms of accuracy and computational efficiency. Preliminary results show that downscaling with a task-specific extension of the foundation model AtmoRep can improve the downscaled product in terms of standard evaluation metrics such as the RMSE compared to a task-specific deep learning model. However, deficiencies in the spatial variability of the downscaled product are also revealed, highlighting the need for future work to focus especially on target data that inhibit a high degree of spatial variability and intrinsic uncertainty such as precipitation.


Contributing Institute(s):
  1. Jülich Supercomputing Center (JSC)
Research Program(s):
  1. 5111 - Domain-Specific Simulation & Data Life Cycle Labs (SDLs) and Research Groups (POF4-511) (POF4-511)
  2. MAELSTROM - MAchinE Learning for Scalable meTeoROlogy and cliMate (955513) (955513)
  3. Verbundprojekt: MAELSTROM - Skalierbarkeit von Anwendungen des Maschinellen Lernens in den Bereichen Wetter und Klimawissenschaften für das zukünftige Supercomputing (16HPC029) (16HPC029)
  4. Earth System Data Exploration (ESDE) (ESDE)

Appears in the scientific report 2024
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Dokumenttypen > Präsentationen > Poster
Workflowsammlungen > Öffentliche Einträge
Institutssammlungen > JSC
Publikationsdatenbank
Open Access

 Datensatz erzeugt am 2025-01-09, letzte Änderung am 2025-02-03


OpenAccess:
Volltext herunterladen PDF
Dieses Dokument bewerten:

Rate this document:
1
2
3
 
(Bisher nicht rezensiert)