Preprints
https://doi.org/10.5194/egusphere-2025-4531
https://doi.org/10.5194/egusphere-2025-4531
02 Oct 2025
 | 02 Oct 2025
Status: this preprint is open for discussion and under review for Nonlinear Processes in Geophysics (NPG).

Bayesian data selection to quantify the value of data for landslide runout calibration

V Mithlesh Kumar, Anil Yildiz, and Julia Kowalski

Abstract. The reliability of physics-based landslide runout models depends on the effective calibration of its parameters, which are often conceptual and cannot be physically measured. Bayesian methods offer a robust framework to incorporate uncertainties in both model and observations into the calibration process. Therefore, they are increasingly used to calibrate physics-based landslide runout models. However, the practical application of Bayesian methods to real-world landslide events depends on the availability and quality of observational data, which determines the reliability of the calibration outcomes. Despite this, systematic investigation of the influence of observational data on the Bayesian calibration of landslide runout models has been limited.

We propose quantifying the impact of observational data on calibration outcomes by measuring the information gained during the calibration process using a decision-theoretic measure called Kullback-Leibler (KL) divergence. Building on this, we present a unified Bayesian data selection workflow to identify the most informative dataset for calibrating a given parameter. The workflow runs parallel calibration routines across available observation datasets. It then computes the information gained relative to the observations by calculating the KL divergence between prior and posterior distributions and selects the dataset that yields the highest KL divergence.

We demonstrate our workflow using an elementary landslide runout model, calibrating friction parameters with a diverse set of synthetic observations to evaluate the impact of data selection on parameter calibration. Specifically, we compare and quantify the information gained from calibration routines using observations with varying information content, i.e., velocity vs. position, and observations with different granularity, i.e., aggregated data vs. time series data. The insights from this study will optimize the use of available observations for calibration and guide the design of effective data acquisition strategies.

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.
Share
V Mithlesh Kumar, Anil Yildiz, and Julia Kowalski

Status: open (until 27 Nov 2025)

Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor | : Report abuse
V Mithlesh Kumar, Anil Yildiz, and Julia Kowalski

Data sets

Bayesian data selection to quantify the value of data for landslide runout calibration V Mithlesh Kumar https://doi.org/10.5281/zenodo.17120721

Interactive computing environment

Bayesian data selection to quantify the value of data for landslide runout calibration V Mithlesh Kumar https://doi.org/10.5281/zenodo.17120721

V Mithlesh Kumar, Anil Yildiz, and Julia Kowalski
Metrics will be available soon.
Latest update: 02 Oct 2025
Download
Short summary
The reliability of Bayesian calibration depends on the quality and availability of observational data. But are we choosing the right data? We address this question by measuring the information gained during calibration to quantify how data selection influences the Bayesian calibration of physics-based landslide runout models. We find that more data does not always yield better results – observations that capture the dynamics governed by a parameter are more effective for its calibration.
Share