the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Technical Note: Can Visual Gauges Trained on Biased Contact-based Gauge Data Accurately Estimate River Stage?
Abstract. Water stage variations significantly influence biochemical and hydrological processes within river networks. River camera, with its ease of deployment and low cost, has emerged as a promising tool for water stage estimation, enabling efficient water stage interpretation from images via deep learning (DL). However, a critical challenge is the requirement of accurate water stage data for DL training, which often have biases caused by sedimentations, floating debris or water flow impacts associated with contact-based gauge observations. Previous studies have overlooked the influence of gauge data errors in real-world applications. This study introduces an imaging-based water stage estimation framework that addresses hidden errors in gauge station measurements for training DL models. The framework adopts a multi-task learning paradigm, using erroneous gauge stage data as labels and incorporating water pixel ratios automatically extracted from images to constrain model estimation ranking. Based on training loss, a thresholding method then filters error-free data to retrain an unbiased model. This framework is tested on images and bubble-gauge stage data from the Minturn River, Greenland, spanning 2019 to 2021. The results obtained show the framework successfully identified a gauge offset event on July 29, 2021, and mitigated an average water stage observation error of approximately 0.6 meters thereafter. Moreover, the trained DL model revealed water stage fluctuations under low-flow conditions that gauge observation could not reflect. This study implies that integrating contact and non-contact observations is a robust approach for river stage measurement.
- Preprint
(1789 KB) - Metadata XML
- BibTeX
- EndNote
Status: open (until 18 Dec 2025)
-
RC1: 'Comment on egusphere-2025-2902', Anonymous Referee #1, 16 Oct 2025
reply
-
AC1: 'Reply on RC1', Ze Wang, 13 Nov 2025
reply
We sincerely appreciate the reviewer’s insightful comments on our manuscript. Our detailed responses are provided in the attached document.
-
AC1: 'Reply on RC1', Ze Wang, 13 Nov 2025
reply
-
RC2: 'Comment on egusphere-2025-2902', Daniel Glaser, 03 Dec 2025
reply
The authors present a methodology in which they use river camera footage for sensing river stage remotely. They apply an algorithm that computes the percentage of pixels depicting water surface and pair this metric with stage data from a bubble-gauge to train a deep learning regression model. In a second step the difference between model and gauge is tested to see if there is a gauge failure event in the training data and when exactly it took place. The model is then retrained without the biased data.
The methodology was applied to a dataset of the Minturn river in Greenland, covering more than 2 years (2019-2021) and which includes a major gauge failure event in 2021. They are able to obtain a good model performance for the time before gauge failure. With the aid of the afore mentioned test they were able to identify the exact time of failure. The model was then retrained without the biased data, which preserved its functionality, making this potentially a valuable insight.Despite it's apparent scientific value, there are a few issues with the manuscript that make it unsuitable in it's current form. While the overall presentation quality is very good, there are some shortcomings regarding the significance and scientific quality that need to be addressed before publication.
Even though the work is well motivated, the method fails to convince, that it can be generalised. It relies on gauging data for training, which rules out ungauged streams. Furthermore the authors base this methodology on a case study using just a single camera at one location and including only one (very pronounced) gauge failure event. No claim can be made whether this method holds up under different conditions. Finally, the overall accuracy is not able to compete with the current state-of-the-art sensors.
In some parts the scientific quality is hard to judge, since the individual steps are only documented partially. The segmentation process used to obtain the water pixel percentage is not further described, despite it being one of the two input variables for training the model. Even though the process is adopted from a previous publication by some of the authors, it should be described in more detail, as it's a major potential source of errors. The same applies to the model training and retraining process. The authors report little on what exactly was done and to which effect, giving a bare minimum of information and providing no supplementary information.
Nevertheless, I would like to acknowledge the potential scientific value of this work and thank the authors for their well-structured manuscript. I encourage them to address the mentioned gaps by providing further details of their work.
Please find attached a document with further line-by-line comments (major, minor, trivialities).
Note to the editor: due to my limit experience in that field, I cannot adequately judge the scientific soundness of the described deep learning training beyond common sense.
-
AC2: 'Reply on RC2', Ze Wang, 09 Dec 2025
reply
We sincerely thank the reviewer for the insightful and constructive comments on our manuscript. Our detailed responses are provided in the attached document.
-
AC2: 'Reply on RC2', Ze Wang, 09 Dec 2025
reply
Viewed
| HTML | XML | Total | BibTeX | EndNote | |
|---|---|---|---|---|---|
| 2,646 | 65 | 23 | 2,734 | 47 | 43 |
- HTML: 2,646
- PDF: 65
- XML: 23
- Total: 2,734
- BibTeX: 47
- EndNote: 43
Viewed (geographical distribution)
| Country | # | Views | % |
|---|
| Total: | 0 |
| HTML: | 0 |
| PDF: | 0 |
| XML: | 0 |
- 1
The authors present a technical note introducing an AI-based approach for river stage measurement using camera imagery, leveraging a multi-task learning framework. The core idea is to directly learn stage information from images while incorporating relative stage information from an image segmentation task as an auxiliary loss. The consideration of a multi-task learning approach is interesting, as it offers a potentially more robust way to train the neural network, and the effort to automatically filter noisy stage reference data is noted. While the study explores a novel application of multi-task learning, I think, several fundamental and methodological issues prevent its suitability for publication in HESS in its current form.
The most critical concern lies in the overall suitability and motivation for an approach that learns the absolute stage directly from images. The reliance on on-site gauge data for training at every new location significantly limits its utility, particularly for ungauged catchments, which are the primary target for innovative remote sensing techniques. As gauged catchments already possess well-established, high-accuracy stage measurement methods, the practical added value of this camera-based approach for these sites is questionable. Also, there are already studies discussing the potential and limits of directly learning the stage from images, which are not mentioned in this study (e.g. Vanden Boomen et al., 2021). Furthermore, the risk is high that the approach is highly sensitive to any movements (internal or external geometry) of the camera setup. Such movements would likely necessitate a complete re-learning of the model, which is a significant practical limitation and is not adequately addressed in the current work. Finally, the authors' premise that obtaining accurate stage data is a critical challenge for all DL-based camera gauges is debatable. For approaches relying on photogrammetry, the stage data serves only as a reference, not as the primary input for the AI-model, thereby mitigating this "critical challenge." A stronger, more refined motivation for this specific DL-only approach is needed.
The paper utilizes pixel information from segmented images to provide relative stage information but lacks sufficient discussion on the segmentation process itself. This is a significant omission, especially since several established studies (e.g., Eltner et al., 2021; Zamboni et al., 2025, Moghimi et al., 2024) already perform this kind of water segmentation for stage measurement, and the potential for segmentation errors and their influence on the multi-task learning is not discussed at all. Furthermore, the review fails to include relevant, state-of-the-art photogrammetric approaches that use water segmentation (e.g., Blanch et al., 2025). Given that the study site appears highly suitable for these methods, a direct comparison and justification for choosing the DL-only approach is necessary. Also, the achieved accuracy, appearing to be in the decimeter (dm) range, is not competitive with the centimeter (cm) accuracy demonstrated by other camera gauge studies, particularly those using robust photogrammetric methods (e.g., Eltner et al., 2021, Erfani et al., 2023, Blanch et al., 2025). Therefore, also the title of the manuscript is misleading because I think, the achieved accuracies cannot be described accurate. Finally, the approach involves combining two loss functions, which necessitates the fine-tuning of the lambda value. This introduces a hyperparameter that must be manually tuned, complicating the model's reliability and generality.
The suggested automatic detection of gauge errors appears effective only for very strong and obvious errors. It is unclear why an established statistical approach would not be equally or more effective for this task. The authors apply an automatic post-processing/filtering step to refine the training data, assuming the error resides in the stage data and not the camera imagery. This assumption needs stronger justification. The lack of provided code is a serious concern, particularly for a technical note. This does not comply with the FAIR principles, which are essential for research reproducibility.
While the multi-task learning idea is technically interesting, the manuscript does not provide a compelling scientific or practical justification for an approach that learns stage directly from images given the high site-specificity, sensitivity to movement, and lower accuracy compared to established methods. The fundamental questions regarding transferability and the need for new methods at gauged catchments remain unanswered. Furthermore, significant methodological detail is missing, and the paper does not adhere to open science principles.
References:
Blanch, X., Grundmann, J., Hedel, R., & Eltner, A. (2025). AI Image-based method for a robust automatic real-time water level monitoring: A long-term application case. https://doi.org/10.5194/egusphere-2025-724
Eltner, A., Bressan, P. O., Akiyama, T., Gonçalves, W. N., & Marcato Junior, J. (2021). Using Deep Learning for Automatic Water Stage Measurements. Water Resources Research, 57(3). https://doi.org/10.1029/2020WR027608
Erfani, S. M. H., C. Smith, Z. Wu, E. A. Shamsabadi, F. Khatami, A. R. J. Downey, J. Imran, and E. Goharian. 2023. “Eye of Horus: A Vision-Based Framework for Real-Time Water Level Measurement.” Hydrology and Earth System Sciences 27 (22): 4135–4149. https://doi.org/10.5194/ hess-27-4135-2023 .
Moghimi, A., M. Welzel, T. Celik, and T. Schlurmann. 2024. “A ComparativePerformance Analysis of Popular Deep Learning Models and Segment Anything Model (SAM) for River Water Segmentation in Close-Range Remote Sensing Imagery.” Institute of Electrical and Electronics Engineers Access 12:52067–52085. https://doi.org/10.1109/ACCESS.2024.3385425
Vanden Boomen, R. L., Yu, Z., & Liao, Q. (2021). Application of Deep Learning for Imaging-Based Stream Gaging. Water Resources Research, 57(11). https://doi.org/10.1029/2021WR029980
Zamboni, P. A. P., Blanch, X., Marcato Junior, J., Gonçalves, W. N., & Eltner, A. (2025). Do we need to label large datasets for river water segmentation? Benchmark and stage estimation with minimum to non-labeled image time series. International Journal of Remote Sensing, 46(7), 2719–2747. https://doi.org/10.1080/01431161.2025.2457131