English
 
Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Deep reinforcement learning of model error corrections

Authors

Finn,  Tobias
IUGG 2023, General Assemblies, 1 General, International Union of Geodesy and Geophysics (IUGG), External Organizations;

Durand,  Charlotte
IUGG 2023, General Assemblies, 1 General, International Union of Geodesy and Geophysics (IUGG), External Organizations;

Farchi,  Alban
IUGG 2023, General Assemblies, 1 General, International Union of Geodesy and Geophysics (IUGG), External Organizations;

Bocquet,  Marc
IUGG 2023, General Assemblies, 1 General, International Union of Geodesy and Geophysics (IUGG), External Organizations;

External Ressource
No external resources are shared
Fulltext (public)
There are no public fulltexts stored in GFZpublic
Supplementary Material (public)
There is no public supplementary material available
Citation

Finn, T., Durand, C., Farchi, A., Bocquet, M. (2023): Deep reinforcement learning of model error corrections, XXVIII General Assembly of the International Union of Geodesy and Geophysics (IUGG) (Berlin 2023).
https://doi.org/10.57757/IUGG23-3336


Cite as: https://gfzpublic.gfz-potsdam.de/pubman/item/item_5019665
Abstract
Deep reinforcement learning has empowered recent advances in games like chess or in language modelling with chatGPT, but can also control nuclear fusion in a Tokamak reactor. In the often-used actor-critic framework, a neural network is trained to control while another neural network evaluates the actions of the first network. In this talk, we cast model error correction into a remarkably similar framework to learn from temporally sparse observations. A first neural network corrects model errors, while a second, simultaneously trained, estimates the future costs if the model error correction were applied. This allows us to circumvent the need for the model adjoint or any linear approximation for learning in a gradient-based optimization framework.We test this novel framework on low-order Lorenz and sea-ice models. Trained on already existing trajectories, the actor-critic framework can not only correct persisting model errors, but significantly surpasses linear and ensemble methods. Furthermore, using this framework enables us to learn unknown processes in sea-ice models from temporally sparse observations. Therefore, we see training of model error corrections with such an actor-critic framework as one of the most promising steps for geoscientific hybrid modelling.