Climate Change Adaptation Digital Twin: a window to the future of our Planet 

The Climate Change Adaptation Digital Twin of the EU Commission's Destination Earth initiative (Climate DT) is a step change in the way information on climate change is provided, aiming to help face the complex challenges posed by our changing climate. The Climate DT represents a pioneering effort to produce operationally multi-decadal climate projections. It leverages the latest scientific and technology advances, including the unprecedented computing power provided by first pre-exascale supercomputers in Europe of the European High Performance Computing Joint Undertaking (EuroHPC JU). The Climate DT will provide regularly updated, globally consistent Earth system and impact sector information at scales of a few kilometres, where many of the impacts of climate change are felt. It will thus reduce the update frequency of the cycles of climate projections from 7-10 years now to yearly or less. At the same time, it is setting a unique capability to produce bespoke cutting edge numerical simulations addressing ‘what-if’ questions related to the impact of certain scenarios or policy decisions on the evolution of our planet. The Climate DT projections  – from global to local scale, for several decades ahead – will enable decision-making in support of climate change adaptation and the implementation of the European Green Deal. 

Let’s take a closer look at this innovative system, developed by a wide partnership consortium led by CSC - IT Center for Science, through a contract procured by the European Centre for Medium-range Weather Forecasts (ECMWF) in the framework of the Destination Earth initiative, funded by the European Commission. This partnership  involves leading climate centres, supercomputing centres, national meteorological services, academia and industrial partners from six European countries.

Recent years, and especially the last months, have shown the extent at which the climate crisis is amplifying. We have entered uncharted territory. The Copernicus Climate Change Service (C3S) ERA5 reanalysis dataset shows that global mean temperature in 2023 was the highest since the start of the records (1940), and we have  experienced a 12-month period with an average global mean temperature above 1.5ºC compared to the pre-industrial era. 

Monthly global surface air temperature anomalies (°C) relative to 1850–1900 from January 1940 to March 2024, plotted as time series for each year. 2024 is shown with a thick yellow line, 2023 with a thick red line, and all other years with thin lines shaded according to the decade, from blue (1940s) to brick red (2020s). Data source: ERA5. Credit: C3S/ECMWF.

Monthly global surface air temperature anomalies (°C) relative to 1850–1900 from January 1940 to March 2024, plotted as time series for each year. 2024 is shown with a thick yellow line, 2023 with a thick red line, and all other years with thin lines shaded according to the decade, from blue (1940s) to brick red (2020s). Data source: ERA5. Credit: C3S/ECMWF.

Monthly global surface air temperature anomalies (°C) relative to 1850–1900 from January 1940 to March 2024, plotted as time series for each year. 2024 is shown with a thick yellow line, 2023 with a thick red line, and all other years with thin lines shaded according to the decade, from blue (1940s) to brick red (2020s). Data source: ERA5. Credit: C3S/ECMWF.

In other words, we are increasingly getting closer to the 1.5ºC global warming limit set in the Paris Agreement. At the same time certain extreme weather events are also becoming more frequent and intense, with often catastrophic impacts. This fast-evolving situation motivates enhancing our capabilities to better respond and adapt to the environmental challenges in a climate change context, beyond what can be currently done with the existing sources of climate information.

This is one of the main goals of the ambitious Destination Earth initiative of the European Commission (DestinE), which aims at developing a digital replica of our planet, with digital twins simulating different aspects of the Earth system. This new type of information system, with enhanced detail, quality, consistency and interactivity, also allows us to address relevant ‘what-if’ questions related to the Earth-system. 

ECMWF and its partners from 90 institutions across Europe implement the first two, high-priority digital twins of DestinE on Weather-induced Extremes and Climate Change Adaptation (Climate DT) as well as the Digital Twin Engine, an innovative software architecture supporting running the digital twins and interacting with them and their data. The delivery of the digital twins critically relies on the strategic partnership with the EuroHPC JU, that provides special access to its supercomputers, since DestinE is considered a strategic initiative for the European Union. This partnership allows to translate Europe's investments in pre-exascale supercomputers into benefits for the society.

Destination Earth’s Climate DT will provide a new capability to determine the impacts of scenarios or policies on the climate system and relevant impact-sectors, at global, regional, and local scales several decades ahead using cutting-edge numerical simulations. 

The Climate DT represents the first effort to operationalize multi-decadal climate projections that will produce globally consistent Earth system and impact-sector information at local scales (of a few km) where many of the impacts of climate change are observed. 

A satellite image of the future: a simulated satellite image for a period of 100-days (1 May to 8 August 2032) derived from one of the prototype projections of the Climate DT with IFS-NEMO (at 4.4km resolution for the atmosphere and land, and 1/12 degree for the ocean and sea-ice), performed on the EuroHPC LUMI supercomputer. Such level of detail obtained now with km scale climate simulations was so far only available from weather prediction models.

A satellite image of the future: a simulated satellite image for a period of 100-days (1 May to 8 August 2032) derived from one of the prototype projections of the Climate DT with IFS-NEMO (at 4.4km resolution for the atmosphere and land, and 1/12 degree for the ocean and sea-ice), performed on the EuroHPC LUMI supercomputer. Such level of detail obtained now with km scale climate simulations was so far only available from weather prediction models.

Importantly, these projections will be updated regularly (at least on a yearly basis, for example using improved models, different scenarios), and on demand to address specific ‘what-if’ questions. This will change the level of support currently available to design adaptation measures by, for example, assessing climate impacts under several policy-relevant scenarios, or creating dedicated storylines of how recent extreme events with high impacts on climate-vulnerable sectors, have been impacted by climate change and may unfold in different future climates. 

The Climate DT is implemented by a strong partnership led by the CSC-IT Center for Science, through a contract procured by ECMWF, which started in November 2022.

The teams involved in the Climate DT have been working intensely in close collaboration with teams at ECMWF to develop the key components of the Climate DT, deploy them on  the EuroHPC LUMI system, the first pre-exascale computer in Europe, and perform the first ever prototype projections at 5km for a couple of decades. 

Stay tuned for more information on the key achievements of the Climate DT during the first phase of DestinE, which ends in June 2024 and the outlook for the second phase of the initiative (June 2024 - June 2026). The European Commission will mark the launch of DestinE with a high-level event held at LUMI data center in Finland on 10 June 2024.

Climate DT: main components and features  

The long-term vision of the Climate DT consists of providing a source of novel high-resolution global climate information for both the present and the future giving end users the possibility to assess the effectiveness of different adaptation strategies. 

The Climate DT is designed for the continuous production of climate data and projections, ensuring that high-quality, up-to-date information is always readily available. It is also designed for on-demand production, allowing to rerun the global models to address specific "what-if" questions, or even to “insert” twins-within-the-twin for specific aspects.

Destination Earth Climate DT surface temperature ICON LUMI HPC

A snapshot of the surface temperature from one of the prototype projections of the Climate DT with ICON (at 5km resolution across atmosphere, land, ocean and sea-ice), performed on the EuroHPC LUMI supercomputer.

A snapshot of the surface temperature from one of the prototype projections of the Climate DT with ICON (at 5km resolution across atmosphere, land, ocean and sea-ice), performed on the EuroHPC LUMI supercomputer.

In practice, the Climate DT relies on: global km-scale climate models and flexible workflows, including directly linked impact models. Key features include an operational configuration that includes monitoring and evaluation, and interactivity elements, allowing to enhance the responsiveness to emerging user needs. The recent breakthroughs in Artificial Intelligence (AI) will allow boosting the interactivity of the Climate DT, in particular through the use of tailored chatbots to ease the access to the newly produced climate information, and the use of Machine Learning models to enhance uncertainty quantification. 

The implication of selected users right from the start of its conception, following DestinE’s co-design strategy is key to ensure that the main recipients obtain the climate information they need.

Km-scale, global storm-resolving eddy-rich climate models

The Climate DT employs a global km-scale modelling approach, characterised by several innovative aspects. The km-scale models aim to improve the representation of critical processes and reduce climate model biases, to enhance the overall trustworthiness of climate projections. They provide consistent climate information with local granularity on a global scale, bridging the gap between large-scale climate projections and local climate impacts and avoiding data gaps and inconsistencies that come with existing regional downscaling efforts.

The Climate DT provides consistent climate information with local granularity on a global scale (at 5km or 10km instead of more than 100 km today), bridging the gap between large-scale climate projections and local climate impacts and avoiding data gaps and inconsistencies that come with existing regional downscaling efforts. Snapshots of the surface temperature from the CMIP6 models, right, and from prototype projections of the Climate DT with IFS-NEMO at 4.4 km resolution on LUMI supercomputer.

The Climate DT provides consistent climate information with local granularity on a global scale (at 5km or 10km instead of more than 100 km today), bridging the gap between large-scale climate projections and local climate impacts and avoiding data gaps and inconsistencies that come with existing regional downscaling efforts. Snapshots of the surface temperature from the CMIP6 models, right, and from prototype projections of the Climate DT with IFS-NEMO (at 4.4 km resolution for the atmosphere and land and 1/12 degree resolution for the ocean and sea-ice), performed on the EuroHPC LUMI supercomputer.

The Climate DT provides consistent climate information with local granularity on a global scale (at 5km or 10km instead of more than 100 km today), bridging the gap between large-scale climate projections and local climate impacts and avoiding data gaps and inconsistencies that come with existing regional downscaling efforts. Snapshots of the surface temperature from the CMIP6 models, right, and from prototype projections of the Climate DT with IFS-NEMO (at 4.4 km resolution for the atmosphere and land and 1/12 degree resolution for the ocean and sea-ice), performed on the EuroHPC LUMI supercomputer.

The models underpinning the Climate DT incorporate atmosphere, land, ocean and sea-ice components. They are building on some of Europe’s leading weather and climate models, which are now adapted to km-scale resolutions in close sync with other initiatives (e.g., the Horizon 2020 nextGEMS and Horizon Europe EERIE projects and the Warm World project in Germany), through a truly European collaborative endeavour.  

  • IFS-NEMO has been developed starting from the most recent version of the ECMWF Integrated Forecasting System (IFS) atmospheric/land model  coupled to the latest release of the NEMO model implemented by the Barcelona Supercomputing Center in collaboration with ECMWF.
  • IFS-FESOM also employs most recent version of IFS coupled to FESOM – one of the first global mature sea ice-ocean models formulated on unstructured meshes. Finite-VolumE Sea ice-Ocean Model (FESOM) is a development of the Alfred Wegener Institute (AWI). Read more on the developments of IFS-FESOM towards km-scales in the framework of nextGEMS.
  • ICON, a flexible, scalable, high-performance modelling framework for weather, climate and environmental prediction. The ICOsahedral Non-hydrostatic model has been jointly developed by the German Weather Service (DWD), Max-Planck-Institut für Meteorologie, (MPI-M), Deutsches Klimarechenzentrum (DKRZ), Karlsruhe Institute of Technology (KIT) and the Center for Climate Systems Modeling (C2SM).
High-resolution simulations of surface winds, sea ice and ocean currents using the models involved in the Climate DT (ICON, IFS-FESOM and IFS-NEMO). Visualisations created in the framework of DestinE and of Horizon 2020 nextGEMS project. 

High-resolution simulations of surface winds, sea ice and ocean currents using the models involved in the Climate DT (ICON, IFS-FESOM and IFS-NEMO). Visualisations created in the framework of DestinE and of Horizon 2020 nextGEMS project, using simulations at resolutions ranging between 5 and 10 km for the various earth system components.

High-resolution simulations of surface winds, sea ice and ocean currents using the models involved in the Climate DT (ICON, IFS-FESOM and IFS-NEMO). Visualisations created in the framework of DestinE and of Horizon 2020 nextGEMS project, using simulations at resolutions ranging between 5 and 10 km for the various earth system components.

The Climate DT leverages some of the world's most powerful pre-exascale HPC systems, those of the EuroHPC JU, a task that presents significant challenges but offers unparalleled opportunities (e.g., exploiting the km-scale in multi-decadal climate projections). This involves a continuous substantial effort to adapt and optimise the climate models to the pre-exascale European HPC platforms, to make efficient use of heterogeneous architectures, using both general-purpose and accelerated processors, and the corresponding data-handling capacity. This is an absolute requirement to efficiently perform kilometre-scale global simulations (with a target of 5 km) on a multi-decadal timescale on HPC systems distributed across Europe.

Operational projections using flexible workflows

Developing a digital twin of the complex Earth-system delivering operational multi-decadal projections requires strictly following best practices in software engineering. 

These include the generalised use of continuous integration and delivery, agile software development methodologies, best code structuring and documentation practices, performance analysis and performance portability. 

Autosubmit, the workflow manager solution used in the Climate DT on the EuroHPC platforms, can create and manage workflows, which orchestrate different elements of the Climate DT and allows to incorporate an arbitrary number of data consumers. 

The Climate DT employs flexible workflows for the operational (regular) production of climate projections, embedding selected applications. The length of CMIP (Coupled Model Intercomparison Project) cycles is significantly reduced, from 7–10 years to just a year and less, enhancing the responsiveness to emerging policy requirements and user needs.

In times of transformative change in the field of climate science, such as the advent of AI, short-update cycles are crucial for incorporating the latest advancements into the production of the best possible climate information. 

The regular production entails scenario projections for the period 2020–2050 following the emission scenarios defined by ScenarioMIP. Historical simulations starting in 1990 following a modified HighResMIP protocol, as well as long control simulations with constant 1950 forcing are used to evaluate model performance.

An important part of the operationalisation of multi-decadal projections is defining a common data portfolio and data governance following international standards. This implies providing the same parameters, in the same format, for all models, and all simulations, which is an important aspect of making the data usable. This ensures that once a user has adapted to the DestinE data formats, they can use new simulations from any of the participating models straightaway without further adjustments. 

The simulation outputs as defined in the DestinE portofolio are transferred to the DestinE Data Lake, implemented by EUMETSAT and can be accessed via the DestinE Core Platform implemented by ESA. The data portfolio is not static and will be enriched with new variables and frequencies of interest according to the user requirements that are continuously being identified and updated.

BSC team launching one of the Climate DT simulations and monitoring its progress with the Autosubmit workflow managers. Credit: BSC

BSC team launching one of the Climate DT simulations and monitoring its progress with the Autosubmit workflow managers. Credit: BSC

BSC team launching one of the Climate DT simulations and monitoring its progress with the Autosubmit workflow managers. Credit: BSC

Quality assurance, evaluation and uncertainty quantification

Another important aspect is to ensure that quality control measures are in place to continuously monitor the accuracy and reliability of the climate data produced. This implies applying a battery of tests to ensure consistency and completeness before making the data accessible to users.

It also implies a rigorous scientific evaluation of the climate simulations, performed in two phases. First, an exhaustive monitoring of the simulations, compared with observations (for historical simulations) or benchmark simulations (for projections) to assess their realism. This includes innovative diagnostics that make use of the full model output. For example, ocean eddies and tropical storms can be analysed with unprecedented detail given the wealth of variables, frequencies, and resolution available. In a second phase, the diagnostics focus on a deeper analysis documenting evaluation results relevant for climate adaptation.

This evaluation informs the evolution of the climate DT by identifying problems to be addressed and improved by model development. This continuous evolution process ensures that the climate DT is providing high-quality information fit for evolving user needs.

The climate DT combines this comprehensive quality assessment with a quantification of the main sources of uncertainty. Uncertainty quantification is needed to provide essential context to translate the information produced by the climate DT into policy decisions, adaptation strategies, and risk assessment for business plans. 

Climate DT (here IFS-NEMO) performance indices compared to a subset of CMIP6 models. Values below 1 (blue colors) indicate that the simulations is closer to observations than average CMIP6 models. The performance indices provide a representative coverage of parameters and regions. 

Climate DT performance indices compared to a subset of CMIP6 models. Values below 1 (blue colors) indicate that the simulation (here an historical run with IFS-NEMO at 9km resolution for atmosphere and land, and 1/12 deg for ocean and sea-ice) is closer to observations than the average of CMIP6 models. The performance indices provide a representative coverage of parameters and regions.

Climate DT performance indices compared to a subset of CMIP6 models. Values below 1 (blue colors) indicate that the simulation (here an historical run with IFS-NEMO at 9km resolution for atmosphere and land, and 1/12 deg for ocean and sea-ice) is closer to observations than the average of CMIP6 models. The performance indices provide a representative coverage of parameters and regions.

Streaming earth system information to applications

The Climate DT introduces a paradigm shift in the generation of information for climate adaptation. Most of the current climate information sources, including global and regional climate projections, lack flexibility to adapt the outputs to the users’ requirements. Moreover, the large model output volumes prevent storing the full information content. So, radically new ways need to be devised.

This motivates one of the key innovations of the Climate DT to address these challenges: the streaming approach. The streaming of the full model output, with all the variables available, at native spatial resolution and the highest frequency possible, gives users immediate access (on the HPC) to the complete set of climate data generated by the models.

Moreover, the Climate DT produces data in a standardised format  across different models and model components (so-called Generic State Vector) and a unified grid equal-area hierarchical HealPix grid, optimised for hierarchical data exploitation as is required in high-resolution visualisation, local data retrieval and AI/ML training. This approach is already quickly gaining traction in the community for efficiently working with very high resolution data.

The Digital Twin on Climate Change Adaptation, Sebastian Milinski

From co-design to co-production of data

Selected applications were embedded in the Climate DT right from the start, through a co-production approach. 

The co-production, an iterative process, allows the Climate DT team to deepen their understanding of the user needs and to explore with users what can be done better with the novel information. This allows to better shape the digital twin and optimise the experimental set up of the continuous simulations. This process ensures the relevance of the data produced for the climate adaptation community.

New impact models and new sets of indicators that estimate climate-related variables relevant for key sectors such as renewable energies, water management, or agriculture can be incorporated between different simulations. 

This strategy is also applied to develop bias-adjusted data directly from the model simulations for those impact models that require it. 

The bias adjustment parameters are estimated and continuously refined using the model output from previous simulations. In this online context, this means that the impact model and the bias adjustment can be run simultaneously with the global climate model. 

Climate Digital Twin Use case wind BSC

The Climate DT data can for example be used to inform the investment strategies of the wind power sector. Different wind energy relevant indicators will be computed in one of the use cases included in the Climate DT, including hourly wind speed distribution with its changes at the multi-decadal scale to help improve wind farm design, and wind turbine resources and vulnerability. Illustrated here is capacity factor for a class S wind turbine (Vestas V164 - 9.5 MW), from one of the prototype runs of the Climate DT with IFS-NEMO (at 4.4.km resolution for atmosphere and land, and 1/12 degree for ocean and sea-ice), performed on EuroHPC LUMI supercomputer.

The Climate DT data can for example be used to inform the investment strategies of the wind power sector. Different wind energy relevant indicators will be computed in one of the use cases included in the Climate DT, including hourly wind speed distribution with its changes at the multi-decadal scale to help improve wind farm design, and wind turbine resources and vulnerability. Illustrated here is capacity factor for a class S wind turbine (Vestas V164 - 9.5 MW), from one of the prototype runs of the Climate DT with IFS-NEMO (at 4.4.km resolution for atmosphere and land, and 1/12 degree for ocean and sea-ice), performed on EuroHPC LUMI supercomputer.

Why a Climate Change Adaptation Digital Twin?

In the face of rapid climate change there is much yet to be known about climate change at the regional and local level, as highlighted by the Intergovernmental Panel on Climate Change (IPCC) in 2021.  

The limited confidence in key aspects of how the climate is changing at regional and local scales shows a need for improved assessment of climate-related risks associated with both local and global processes, and in particular to integrate rapidly changing and unexpected risks or wildcards. In this context, climate information sources need to be timely and driven by societally-relevant questions, something that current practice, mainly driven by research efforts and curiosity, is not able to satisfy. 

The current global climate models are not able to explicitly represent and resolve a number of key processes like ocean eddies and deep atmospheric convection, nor their impact on the global circulation. This results in inaccurate or incomplete representations of key regional aspects which in turn, tends to impact the regional models used to adapt to regional and local changes in our climate. 

Bridging the gap requires new ways to generate climate information that go beyond traditional climate modelling and impact assessment. Estimating the changes in climate and its extremes reliably requires better models that take advantage of the increasingly richer amount of observations, computing, and scientific knowledge of physical processes. These new global models must credibly simulate climate on regional and local scales taking full account of both the remote and local processes that matter for the adaptation of climate-sensitive sectors.

Besides, the availability of increasingly powerful high-performance computing (HPC) systems and recent developments in model efficiency and scalability make it possible to perform climate simulations several decades ahead at resolutions where critical small-scale processes such as ocean eddies atmospheric convection start to be resolved. 

Several groups worldwide are already developing successful prototypes of eddy-rich, storm-resolving climate models with a grid-spacing of less than 10 km. This provides more realistic simulations by reducing the dependence on subgrid parametrisations and better resolving the Earth’s topography. Most of these endeavours are research projects such as H2020 nextGEMS and Horizon Europe EERIE.

Ocean currents, sea ice and snow in the nextGEMS cycle 3 simulation with ICON at 5km , Credit: MPI/DKRZ 

A digital twin of the climate system allows to simultaneously achieve better global models and produce interactive information for climate adaptation. It makes use of observations and integrates a climate model and impact-sector models for vulnerable sectors like energy resource management, all included into a technological solution that favours interaction. The twin allows the assessment of changes and their causes consistently across both local and global spatial scales and a range of time scales. 

The co-design of digital twins with users from the impact-sectors most affected by climate change breaks the current paradigm of climate models with fixed and static flows of information. This approach shifts the provision of climate information for adaptation from the current sources that are often static, poorly funded and inadequate to react to user requirements in a more dynamic, operational and user-relevant way. 

Furthermore, allowing the transformation of highly specialised climate data into user-relevant climate information and shifting the climate information generation to a transdisciplinary production mode are other key motivations for building a digital twin for climate change adaptation. 

Return to the news page

Further information

Destination Earth is a European Union funded initiative launched in 2022, with the aim to build a digital replica of the Earth system by 2030. The initiative is being jointly implemented by three entrusted entities: the European Centre for Medium-Range Weather Forecasts (ECMWF) responsible for the creation of the first two ‘digital twins’ and the ‘Digital Twin Engine’, the European Space Agency (ESA) responsible for building the ‘Core Service Platform’, and the European Organisation for the Exploitation of Meteorological Satellites (EUMETSAT), responsible for the creation of the ‘Data Lake’.   

The Climate DT, procured by ECMWF is developed through a contract led by CSC-IT Center for Science and includes: Alfred Wegener Institute Helmholtz Centre for Polar and Marine Research (AWI), Barcelona Supercomputing Center (BSC), Max Planck Institute for Meteorology (MPI-M), Institute of Atmospheric Sciences and Climate (CNR-ISAC), German Climate Computing Centre (DKRZ), National Meteorological Service of Germany (DWD), Finnish Meteorological Institute (FMI), Hewlett Packard Enterprise (HPE), Polytechnic University of Turin (POLITO), Catholic University of Louvain (UCL), Helmholtz Centre for Environmental Research (UFZ) and University of Helsinki (UH).

We acknowledge the EuroHPC Joint Undertaking for awarding this project strategic access to the EuroHPC supercomputers LUMI, hosted by CSC (Finland) and the LUMI consortium, Marenostrum5, hosted by BSC (Spain) Leonardo, hosted by Cineca (Italy) and MeluXina, hosted by LuxProvide (Luxembourg) through a EuroHPC Special Access call. 

More information about Destination Earth is on the Destination Earth website and the EU Commission website.  

For more information about ECMWF’s role visit ecmwf.int/DestinE   

For any questions related to the role of ECMWF in Destination Earth, please use the following email links:   

General enquiries  

Press and Communications enquiries