Articles | Volume 30, issue 3
https://doi.org/10.5194/hess-30-797-2026
https://doi.org/10.5194/hess-30-797-2026
Research article
 | 
12 Feb 2026
Research article |  | 12 Feb 2026

AI image-based method for a robust automatic real-time water level monitoring: a long-term application case

Xabier Blanch, Jens Grundmann, Ralf Hedel, and Anette Eltner
Abstract

The study presents a robust, automated camera gauge for long-term river water level monitoring operating in near real-time. The system employs artificial intelligence (AI) for the image-based segmentation of water bodies and the identification of ground control points (GCPs), combined with photogrammetric techniques, to determine water levels from surveillance camera data acquired every 15 min. The method was tested at four locations over a period of more than 2.5 years. During this period almost 218 000 images were processed. The results demonstrate a high performance, with mean absolute errors ranging from 0.96 to 2.66 cm in comparison to official gauge references. The camera gauge demonstrates resilience to adverse weather and lighting conditions, achieving an image utilisation rate of above 95 % throughout the entire period. The integration of infrared illumination enabled 24/7 monitoring capabilities. Key factors influencing absolute error were identified as camera calibration, GCP stability, and vegetation changes. The low-cost, non-invasive approach advances hydrological monitoring capabilities, particularly for flood detection and mitigation in ungauged or remote areas, enhancing image-based techniques for robust, long-term environmental monitoring with frequent, near real-time updates.

Share
1 Introduction

The use of image-based systems has transformed the field of geosciences, offering precise and efficient tools for the monitoring and analysis of environmental phenomena. The integration of cameras and photogrammetry in geoscientific research enables the continuous collection of real-time data, facilitating the study of dynamic processes and the acquisition of detailed information on changes in landscapes and ecosystems. These observation systems have been demonstrated to be particularly beneficial in the monitoring of rivers (Eltner et al., 2018; Manfreda et al., 2024), rock and glacier landscapes (Blanch et al., 2023a; Ioli et al., 2023), soil surface (Epple et al., 2025) and vegetation evolution (Iglhaut et al., 2019) among others. They offer a robust and less intrusive alternative to traditional methods, and their low cost and straightforward implementation (Blanch et al., 2024) make them suitable for deployment in remote or less developed areas, thereby expanding the scope of monitored elements and reducing vulnerability to natural disasters.

Monitoring river water levels is a basic but fundamental metric for understanding river behaviour and having this information in real time is crucial for managing flood risks. The ability to detect, predict and mitigate the consequences of water level changes is immensely useful for disaster managers and for minimizing the impacts to the community. Conventionally, obtaining this metric has relied on established gauging stations equipped with contact sensors, such as float gauges and pressure transducers (Morgenschweis, 2018), or active remote sensors like radar and ultrasonic devices (Herschy, 2008). While these established methods provide high-precision data and constitute the backbone of hydrological networks, they face significant operational limitations. Contact sensors are intrusive and prone to damage during high-flow events, while both acoustic and radar systems entail high installation and maintenance costs, limiting their deployment density in remote or financially constrained catchments. Although initiatives have developed low-cost sensor networks based on ultrasonic water level measurements for rivers and urban flash-flood monitoring (Bartos et al., 2018; Mydlarz et al., 2024), these point-based sensors still lack the ability to visually verify the measurements or to capture the surrounding flood dynamics. Therefore, the development of low-cost, automated water level detection systems based on camera gauges is highly desirable, as they can provide accurate and continuous data while also offering visual context, thereby supporting early warning systems and significantly improving the response to critical events (Manfreda et al., 2024).

In particular, the utilization of image-based systems for river monitoring offers a number of advantages over the use of conventional gauging stations. These include greater flexibility in camera placement, the ability to monitor multiple points of a river simultaneously, and reduced costs associated with system installation and maintenance. Additionally, cameras allow for data acquisition in adverse conditions and at very short time intervals, providing a comprehensive and uninterrupted perspective of riverine fluctuations. These advantages make image-based systems an optimal tool for the management and study of water resources – encompassing, for example, turbidity assessment (Miglino et al., 2025; Zhou et al., 2024), water-discharge estimation (Eltner et al., 2020), and flood detection (Fernandes et al., 2022), – thereby increasing the capacity to respond to extreme events and facilitating evidence-based decision-making in water management.

While camera gauges have shown promise, their long-term reliability and performance under varied environmental conditions remain challenging. Basic optical-based methods often struggle with long-term continuous operation or automation, particularly during adverse weather conditions or at night. The literature discusses various automatic methods for detecting water levels from images. For instance, one approach is to install scale bars in the observation area for an automatic measurement based on estimating the contact of the water with the scale bar (Kuo and Tai, 2022; Pan et al., 2018), Other methods detect the waterline on stage boards or vertical rock surfaces using perpendicularly oriented cameras (Leduc et al., 2018; Young et al., 2015). These methods are highly efficient and provide good accuracy, but they require physical intervention in the river to install the scale bar. Moreover, scale bars and sensors installed under bridges or in direct contact with the water are vulnerable to mechanical damage, displacement by debris, and complete loss during flood events – precisely when accurate measurements are most critical. Progressive biofouling from algae and sediment accumulation further degrades visibility of graduations over time, necessitating frequent in-situ maintenance.

Beyond scale-bar methods, other approaches have also attempted to work with the full scene. In this regard, another approach involves transforming the scale bars into landmarks, which are points present in the image with known elevations (e.g., obtained through ground surveys). This approach, known as landmark-based water-level estimation (LBWLE by Vandaele et al., 2021), requires identifying these elements in the image and performing interpolation between the two elevations. The accuracy of the measurements is directly related to the ability to identify known elevation points in the images, and the linear interpolation may not correspond to the actual elevation distribution in the images. Similarly, other works utilize semantic segmentation to detect the continuous water-land boundary (waterline) and map its vertical position to a water level using a pre-calibrated elevation data from LiDAR (Muhadi et al., 2021). Another method that does not require any field installation is the Static Observer Flooding Index (SOFI) method (Moy de Vitry et al., 2019), which detects water level variation based on a direct correlation between the number of pixels segmented as water in each image. This method does not provide direct metric values of the water level but does allow the identification of trends during flooding events (Vandaele et al., 2021).

Other approaches, like the one developed in this article, use the strategy of image-to-geometry registration, which involves reprojecting automatically segmented images into 3D models containing metric real-world information. This method, extensively discussed by Elias et al. (2019) and Eltner et al. (2018), enables the estimation of water levels as real-world elevations by establishing a correspondence between 2D image pixels and their corresponding 3D coordinates in a metrically scaled model of the environment.

The contactless method proposed in this study eliminates the need for any in-stream installation, relying solely on a simple camera setup and ground control points (GCPs) positioned within the field of view. By interpreting the whole scene rather than a narrow scale window, this approach provides multi-point extraction across the segmented waterline, capturing spatially heterogeneous water level dynamics along the river cross-section rather than relying on single-point measurements that may not represent the true behaviour of the water surface. This inherent flexibility makes the system less sensitive to local occlusions or damage and more adaptable to site constraints and changing illumination. Furthermore, even when scale-bar methods are automated, reliable operation depends on optical character recognition (OCR) of small printed digits, which is highly susceptible to direct shadows, sun glare, water splashes, and oblique viewing angles that distort character shapes. By anchoring the entire scene in a georeferenced 3D frame, the method provides absolute water surface elevations and could potentially enable the analysis of complementary dynamic and morphological phenomena across the monitored area – capabilities that would remain unavailable in contact-based single-point instrumentation.

Works based on this approach include Eltner et al. (2021), who laid the groundwork for this study; Zamboni et al. (2025), who aimed to estimate water levels using image-to-geometry and by leveraging deep learning segmentation models that minimize the need for annotated datasets, lowering the effort and the computational cost of the image water segmentation; Erfani et al. (2023) who applied the AI and image-to-geometry approach but for a very short time period (sub-daily); and Krüger et al. (2024) who use a low-cost Raspberry Pi-based camera system a to estimate water levels using the approach developed in this study in a flash flood environment. Additionally, Elias et al. (2019) developed a smartphone application, “Open Water Levels”, that utilizes image-to-geometry registration to enable citizen scientists to capture water level measurements using the smartphone as a measuring device. However, while the aforementioned studies have shown the potential for water level monitoring with cameras, none of them address the operational aspects of the system (i.e., long-term use), as they are limited to specific study areas and short-term observations. Moreover, these approaches still face certain limitations that are well-known in the image-based systems, which primarily concern robustness and adaptability of these methods under challenging environmental conditions and nighttime observations. The work presented here addresses these challenges by meeting the robustness criteria defined by Peña-Haro et al. (2021), as it achieves key properties such as continuous image capture throughout the whole day, applicability across different rivers, and the capacity for near real-time data transmission and processing.

To address these challenges, new research is relying on AI solutions to bring robustness to image processing. Unlike classical methods that rely on hand-crafted features such as texture analysis (Stumpf et al., 2016) or edge detection (Ran et al., 2016; Ridolfi and Manciola, 2018), which are susceptible to noise from reflections and illumination changes, Convolutional Neural Networks (CNNs) automatically learn hierarchical feature representations, enabling the precise pixel-wise classification of water bodies even in cluttered scenes (Eltner et al., 2021; Isikdogan et al., 2017; Muhadi et al., 2021). Consequently, object segmentation using CNNs has become an essential tool in data analysis, representing a significant improvement in results by enhancing the performance of traditional computer vision algorithms (Moghimi et al., 2024). Two recent studies have explored the segmentation of water bodies. Wagner et al. (2023) tested 32 neural networks on the RIWA dataset – a dataset specifically created to segment water for monitoring purposes (Blanch et al., 2023b). Moghimi et al. (2024) evaluated the performance of six modern neural networks across different datasets, including the RIWA dataset. In both works, the U-Net neural network (Ronneberger et al., 2015) was selected as the best-performing model for the RIWA dataset. In Wagner et al. (2023), the UPerNet (Xiao et al., 2018) neural network demonstrated a similar accuracy to U-Net but with a considerable reduction in loss, ensuring higher quality during inference. Similarly, Wang et al. (2024) have recently tried the ResUnet + SAM framework to segment water images (including RIWA dataset) in order to monitor the water level trend in UK rivers.

Beyond standard supervised CNNs, recent research has investigated the utility of foundation models and video-based approaches to reduce training efforts. Zamboni et al. (2025) evaluated Space-Time Correspondence Networks (STCN), which treat image time-series as video frames to propagate labels, and compared them against generic pre-trained models such as the Segment Anything Model (SAM) (Kirillov et al., 2023). Their findings indicate that while zero-shot or generic approaches offer operational convenience by bypassing the need for specific training data, they exhibit a notable loss in metric accuracy compared to supervised methods. Generic models often lack the specific semantic understanding required to distinguish river water from similar textures like wet pavement or sky reflections without fine-tuning. Therefore, in this work we employ advanced in-house trained AI models to robustly segment water bodies in automatically acquired images.

Despite all these recent advances in AI and image-based water level monitoring, a critical gap remains: to our knowledge, no existing image-based contactless method has systematically documented operational reliability over extended periods (> 1 year) across multiple sites under full 24/7 conditions, including adverse weather and nighttime. This research addresses this gap through the KIWA project (Künstliche Intelligenz für die Hochwasserwarnung – Artificial Intelligence for Flood Warning) (Grundmann et al., 2024), which aims to develop AI-driven tools for comprehensive flood monitoring and early warning systems. Within this framework, our research addresses the foundational requirement: accurately quantifying river water levels in a fully automated, contactless manner as a basis for also measuring surface velocity and discharge optically – all derived from a single sensor, the camera.

Rather than simply detecting flood events, our primary objective is to provide centimetre-precision water level measurements under both normal and high-flow regimes, operating robustly in real scenarios with real conditions. Although our approach requires greater technical complexity than simpler methods, this investment is justified because the georeferenced imagery establishes the foundation for future monitoring stages, enabling the derivation of surface velocity (among others by narrowing down the search area, i.e., the water mask) and discharge without additional in-stream instrumentation. By combining AI techniques for water segmentation and ground control point detection with established photogrammetric methods for image-to-geometry registration, the approach presented in this research will enable consistent and accurate water level monitoring over extended observation periods. Advancing beyond initial proof-of-concept testing (Eltner et al., 2018, 2021) this work presents a fully operational, contactless system showing 2.5 years of continuous data validated across multiple rivers and seasons, including adverse weather and nighttime conditions, and benchmarked against co-located official gauging stations.

2 Methods and Materials

Achieving robust, 24/7 water level monitoring without in-stream installations requires integrating multiple disciplines. Field surveys and geodesy establish the metric reference frame, photogrammetry enables contactless measurements through image-to-geometry registration, artificial intelligence automates feature extraction under varying conditions, and software engineering orchestrates real-time processing. The workflow comprises three interdependent stages: establishing a georeferenced 3D site model with ground control points (GCPs) visible in the camera's field of view; automated image processing using AI algorithms for river segmentation and GCP detection; and photogrammetric reprojection of segmented water onto the 3D model to derive metric water levels. This integration enables near-real-time operation without manual intervention, as detailed in the following subsections. Figure 1 illustrates the overall methodological workflow, whereas Fig. B1 in Appendix B provides a detailed algorithm flowchart covering all steps from image acquisition to the final water-level estimation obtained with the optical system (wl_opt) and its comparison with the reference gauge value (wl_ref). All acronyms, including those specific to the project, are defined in Appendix A.

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f01

Figure 1Graphical workflow for obtaining water level from an AI strategy for segmenting and identifying Ground Control Points (GCPs) and image-to-geometry for obtaining metric values of the water level.

Download

2.1 Study area and data retrieval

The methodology was validated using images from four monitoring sites in Saxony, Germany, within the framework of the KIWA project (Grundmann et al., 2024). Three stations – Elbersdorf (ELB) on the Wesenitz River, Großschönau 2 (GRO) on the Mandau River, and Lauenstein 4 (LAU) on the Müglitz River – are co-located with official gauging stations operated by the Staatliche Betriebsgesellschaft für Umwelt und Landwirtschaft (the Saxon state company for environment and agriculture). The fourth site, Neukirch (NEU), is located on the Wesenitz River without a co-located gauge. At the three gauged sites, water level measurements derived from the camera system (wl_opt) are validated against official gauge records (wl_ref). These reference measurements are obtained by float-operated and bubble gauges (redundant measurement systems) and provide the most appropriate benchmark for operational validation because they represent independent, quality-controlled records reflecting accepted hydrometric practice (ISO 18365, 2013) and are used directly by flood services for decision-making. This comparison isolates the performance of the camera-based workflow from site-specific factors (e.g., camera placement, local calibration) and ensures the evaluation aligns with operational monitoring standards.

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f02

Figure 2Geographic distribution and monitoring setup of study sites. Map indicates locations of GRO, ELB and LAU sites. Camera installations (a) and field views (b) at each site. Lower panel shows data acquisition periods.

Table 1Image acquisition properties for each study area.

Download Print Version | Download XLSX

The images are captured using surveillance cameras, which provide an integrated solution for capturing images and videos and transmitting the data remotely to a server (Fig. 2a). Three stations (ELB, LAU, GRO) employ Axis Q1645 LE cameras (1920 × 1080 resolution, 3.75 µm pixel size), while NEU uses the Q1615 Mk III model (1920 × 1080 resolution, 2.90 µm pixel size) (Table 1). All cameras are equipped with zoom lenses (3–10 mm focal length) to optimize field-of-view coverage of the monitored river reach (Fig. 2b). Images are acquired at 15 min intervals (96 images per day) and transmitted automatically to a central server. Acquisition has operated continuously since late 2021, accumulating over 2.5 years of data through June 2024 (Table 1). To enable nighttime monitoring, remotely controlled infrared (IR) illumination was installed at ELB (November 2022) and LAU (May 2023), allowing 24 h observation. At GRO and NEU, nighttime images are captured but lack sufficient illumination for reliable water level estimation.

2.2 3D Modelling and Camera Calibration

Contactless water level measurement depends on establishing a metric reference frame that links 2D image pixels to 3D real-world coordinates. This is achieved through a georeferenced 3D model of the monitored river reach and ground control points (GCPs) visible in both the model and camera images. GCPs serve dual purposes: they enable photogrammetric determination of camera exterior orientation (position and viewing direction) and provide the geometric constraints necessary for an accurate 3D model registration. Proper spatial distribution of GCPs across the image frame ensures robust estimation of camera geometry. GCP coordinates were measured with centimeter accuracy using RTK-GNSS, and 3D models were generated using Structure-from-Motion Multi-View Stereo (SfM-MVS) photogrammetry from terrestrial and UAV imagery where feasible (e.g., Eltner and Sofia, 2020; Smith et al., 2016; Westoby et al., 2012), and the resulting model was georeferenced using both fixed and temporary GCPs measured with RTK-GNSS (Fig. 1a). Standard SfM-MVS reconstruction fails underwater due to refraction at the air-water interface. To address this, refraction correction was applied to submerged areas using PyBathySfM (Dietrich, 2017). When photogrammetric riverbed reconstruction was not feasible – due to high water levels or turbidity – cross-sections were surveyed with RTK-GNSS and interpolated to generate a riverbed mesh. Both approaches yield a gauge zero for the optical system (h_opt0), defined as the mean Z-coordinate of a central riverbed region (e.g., ELB: 4.5 m2). The gauge zero of the optical system (h_opt0) may differ from the gauge zero of the reference gauge data (h_ref0) due to differences in survey origin or riverbed morphology. To ensure direct comparability between datasets, we applied a bias correction to wl_opt by adding a constant offset. This offset was defined as the value that minimizes the mean residual between wl_opt and wl_ref, effectively removing the residual bias over the full observation period. The resulting offsets were small (ELB: 2.0 cm, LAU: 1.0 cm, GRO: 3.0 cm).

Photogrammetric image-to-geometry registration requires accurate interior camera parameters (focal length, principal point, distortion parameters). When feasible, cameras were calibrated before installation by capturing multi-perspective images of a coded calibration chart, with parameters estimated via bundle adjustment in Agisoft Metashape (v2.0.1) incorporating chart target coordinates (e.g., Liebold et al., 2023). When pre-installation calibration was not possible (cameras already mounted), interior parameters were estimated by incorporating fixed-camera images into the SfM bundle adjustment used for 3D modeling. Homologous points between mobile and fixed-camera images enabled parameter estimation, yielding an approximate calibration. LAU and GRO employed chart-based calibration; ELB and NEU required the SfM-based approach.

2.3 Image classification and filtering

Continuous 24/7 monitoring with surveillance cameras introduces the challenge of processing both daylight and low-light imagery, as cameras automatically switch to monochrome mode at night (Fig. 1b). To address this, images are classified based on RGB channel uniformity at diagnostic image locations: images with identical values across all three channels at all test pixels are classified as nocturnal. Without infrared (IR) illumination, nighttime images lack sufficient contrast for reliable river segmentation and are excluded from processing. At sites equipped with IR illumination (ELB, LAU), nighttime images are processed using night-specific segmentation parameters (Sect. 2.4). Daytime images undergo additional quality filtering based on sharpness. A Sobel edge-detection filter (threshold > 0.4) computes the mean edge strength across the image, enabling automatic exclusion of blurred, out-of-focus, or weather-obscured images (e.g., during heavy snowfall or fog). This filtering ensures that only images with sufficient clarity proceed to segmentation and photogrammetric processing.

2.4 AI Segmentation

Accurate water surface segmentation is fundamental to reliable automated water level estimation. Generic pretrained segmentation models (e.g., Segment Anything, SAM) lack the precision needed for reliable water level monitoring, (Zamboni et al., 2025), necessitating the development of domain-specific training. We employ a convolutional neural network (CNN) to segment water pixels from the background (Fig. 1c), based on the UPerNet architecture with a ResNeXt50 backbone (Xiao et al., 2018). This architecture was selected for its demonstrated performance on river segmentation tasks (Wagner et al., 2023).

The model was trained on the RIWA dataset (River Water Segmentation Dataset; Blanch et al., 2023b), which comprises 1163 manually labelled daylight images of water domains captured with diverse sensors (smartphones, drones, DSLR cameras), supplemented with images from WaterNet (Liang et al., 2020) and ADE20K images (Zhou et al., 2019). Part of the images used to create the RIWA dataset were also obtained with the KIWA project cameras (22 images, 1.9 % of the dataset).

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f03

Figure 3Data used for AI-based segmentation. Green areas represent manually segmented images, while orange/pink areas show the results obtained from the model for both (a) daytime and (b) nighttime images.

Download

To optimize model performance, the original RIWA dataset was iteratively refined using Deep Active Learning (Li et al., 2024). This approach identifies images that are poorly segmented by the current model and prioritizes them for manual annotation and inclusion in subsequent training iterations. Through this process, challenging scenarios – such as adverse weather, transparent water, or strong shadows – were systematically incorporated into the training data. The final daytime dataset comprises 1,337 images, of which 145 (11 %) are KIWA images representing diverse locations, weather conditions, and water levels (Fig. 3a). For nighttime monitoring, a separate model – same network – was trained exclusively on infrared (IR) images. Due to the absence of publicly available IR river datasets, this training relied on 146 manually annotated KIWA IR images capturing varied weather and water level conditions (Fig. 3b). Both datasets are augmented using the Albumentation library (Buslaev et al., 2020), to increase training robustness and cross-site transferability. Augmentation is particularly critical for fixed-camera systems to prevent overfitting due to spatial correlations inherent in images from static viewpoints. Geometric transformations (rotation, scaling, flipping) were applied to break spatial dependencies, while pixel-level modifications (brightness, contrast, hue adjustments) were constrained to realistic ranges to preserve natural image appearance. Data augmentation was applied after generating all 512 × 512 patches from each original image in the training dataset, with each original patch producing four augmented versions. The UPerNet network with ResNeXt50 backbone was trained using the FocalLoss cost function and Adam optimizer (initial learning rate: 0.0001, reduced via ReduceLROnPlateau). Training was conducted on an NVIDIA GTX A6000 GPU with a batch size of 30 for 1000 epochs. Model performance was evaluated on an independent test set of 30 images, intentionally biased toward KIWA images (70 %) to prioritize optimization for operational monitoring conditions rather than generalized performance across all river imagery. The nighttime model was evaluated exclusively on KIWA images due to the lack of external IR datasets. The best-performing daytime model achieved 98.9 % pixel-wise accuracy, while the nighttime model reached 99.1 % accuracy, where accuracy represents the proportion of pixels correctly classified as water or non-water. Once trained, the models were deployed operationally to process incoming images every 15 min, automatically generating segmentation masks alongside each captured image (Fig. 3).

2.5 AI GCP Identification

Precise GCP coordinates in the acquired images are essential for an accurate image reprojection. Although cameras are mounted on fixed supports, thermal expansion, vibrations, and physical disturbances can cause subtle positional shifts that alter GCP image coordinates over time. Therefore, GCPs must be detected in every acquired image to maintain geometric consistency throughout the monitoring period (Fig. 1d). Automatic identification of GCPs in images is addressed differently in the literature, for example, using tracking algorithms (e.g., Eltner et al., 2017), feature descriptors like SIFT or SURF (Chureesampant and Susaki, 2014) or geometric shape-fitting (e.g., Maalek and Lichti, 2021). In this research, we employed an AI-based approach using an adapted R-CNN Keypoint detector neural network, specifically retrained on KIWA imagery to identify GCP centers directly. This method, detailed in Blanch et al. (2025a) achieves sub-pixel precision (< 0.5 pixel) without requiring pre- or post-processing and demonstrates robust performance under conditions where traditional feature-based methods fail (e.g., varying illumination, partial occlusions, weathering; Blanch et al., 2025a). When the detector fails to identify one to four GCPs – due to temporary occlusion, extreme weather, or imaging artifacts – missing coordinates are estimated using the KNNImputer algorithm (scikit-learn). This approach leverages temporal coherence by identifying recent images with similar GCP configurations and interpolating missing values from their coordinates. If more than four GCPs cannot be detected, coordinates from the most recent valid image are assigned. This fallback strategy maintains operational continuity while flagging potentially degraded measurements for quality assessment.

2.6 Photogrammetric process

The photogrammetric process transforms the AI-segmented water boundary from image space to real-world coordinates by establishing the geometric relationship between the 2D image plane and the 3D site model (see Elias et al., 2019; Eltner et al., 2018 for detailed explanation) (Fig. 1e). Exterior orientation parameters – defining camera location and rotation – are computed by solving the collinearity equations that relate the detected 2D GCP coordinates to their known 3D world coordinates. This geometric solution enables reprojection of the 3D point cloud onto the image plane, where each pixel can be assigned its corresponding real-world 3D coordinate. To extract water level from the segmented image, the upper contour of the water mask (the river edge opposite to the camera) is intersected with the reprojected point cloud using nearest-neighbour search. The search is spatially constrained to a central cropped region of the point cloud, ensuring that only reliable portions of the 3D model contribute to the measurement.

Water level is derived from the statistical distribution of Z coordinates along the extracted water-edge points. The median Z value represents the water surface elevation (h_opt), which is then converted to water level by referencing to a baseline elevation (Eq. 1):

(1) wl _ opt = h _ opt - h _ opt 0 + offset

The standard deviation of Z coordinates along the intersected points serves as a quality indicator, with higher dispersion suggesting segmentation uncertainty or non-horizontal water behaviour. Validation of the proposed method is conducted at three sites equipped with official gauging stations (ELB, LAU, and GRO), where wl_opt measurements are benchmarked against reference values (wl_ref). Reference measurements represent 15 min averages computed from 5 min sampling intervals, whereas wl_opt values correspond to instantaneous observations. To account for this temporal mismatch and reduce the influence of short-term fluctuations, accuracy assessment is performed using daily-averaged values for both wl_opt and wl_ref. Because the workflow operates fully autonomously on all acquired images, automated quality control is essential to identify and filter erroneous measurements. Two statistical criteria are applied. First, measurements with excessive standard deviation in the Z coordinates of water-edge points are rejected, as high dispersion indicates that the segmentation boundary intersects multiple elevation levels – inconsistent with the expected horizontal water surface. Second, extreme outliers in wl_opt relative to wl_ref are identified using a modified Tukey filter with twice the interquartile range (2 × IQR), removing measurements that deviate significantly from the expected temporal behaviour.

3 Results

Data from the operational automated camera monitoring system between late 2021 to June 2024 was used, with observation periods ranging from 660 d at NEU to 864 d at GRO and 917 and 927 d at ELB and LAU, respectively (Table 2). Over these periods, the cameras acquired 33 382 images at NEU, 47 109 at GRO, 71 787 at ELB and 65 631 at LAU, providing between 660 and 927 d of 15 min (96 images per day) water level observations at the four stations (Table 2). Except for ELB, where technical issues caused data gaps, image acquisition was continuous throughout the observation periods at all locations (Table 2). Quality filtering via Sobel edge detection removed only a small fraction of the acquired images, with rejection rates below 1 % at ELB and LAU and values of 1.2 % and 9.6 % at GRO and NEU, respectively (Table 2). These rejected images are mainly affected by adverse weather conditions such as fog or heavy snowfall, or by low light levels around sunrise and sunset; in addition, high ambient light in the urban environments of NEU and GRO occasionally triggered the surveillance cameras to remain in day mode during the night, producing low-contrast images that failed the Sobel criterion and explaining the higher rejection rates at these sites (Table 2). Among the images passing the Sobel filter, the AI workflow successfully delivered a result for 99.3 %–100 % of the inputs at all stations, after which the standard-deviation filter rejected between 0.1 % and 3.8 % of these results depending on the site (Table 2). At the three gauged stations (ELB, LAU and GRO), the final Tukey outlier filter based on 2 × IQR further removed only 1.2 %–2.6 % of the standard-deviation-accepted measurements (Table 2). The measurements that remain after this Tukey filtering step define the final validation dataset used to compute the water-level error metrics reported in the following subsections. Relative to all acquired images, the overall proportion of valid water-level estimates ranges from 86.9 % at NEU, where a fraction of the images was incorrectly acquired due to urban lights, to 98.1 % at LAU, with 95.3 % and 97.1 % of the images yielding a valid water level at GRO and ELB, respectively (Table 2).

Table 2Image acquisition, filtering and processing statistics at each monitoring station, summarising the number of acquired images, the fractions removed by filters, and the resulting proportion of valid water-level measurements. No Tukey filter is applied at NEU due to the absence of a reference gauge at this site.

Download Print Version | Download XLSX

Table 3Camera focal length and reprojection error of ground control points (in pixels) at the four monitoring stations.

Download Print Version | Download XLSX

Geometric accuracy of the image-to-geometry registration was assessed using the reprojection errors of the ground control points (GCPs) at each station (Table 3). Average reprojection errors range from 4.9 pixels at NEU and 6.3 pixels at LAU to 17.5 pixels at ELB and 29.0 pixels at GRO. The lower mean errors at LAU and NEU reflect the dedicated calibration-camera setups used at these sites, whereas the higher values at ELB and especially at GRO indicate a more challenging imaging geometry and stronger lens distortions. In all cases, the GCPs were deliberately positioned near the image edges, where distortions are most pronounced, so the reported values represent a conservative estimate of the registration quality between the 3D model and the 2D camera images.

Table 4Statistical comparison between wl_opt and wl_ref over the entire observation period, for individual measurements and daily averaged values.

Download Print Version | Download XLSX

Table 4 reports the error statistics between the camera-based water levels (wl_opt) and the gauge records (wl_ref) for the three gauged stations, considering both all individual results and daily averaged values. For individual estimates, mean absolute errors (MAE) are in the centimetre range at all sites, from 1.26 cm at LAU to 2.66 cm at ELB, with GRO in between. With respect to the ±2 cm accuracy threshold commonly used in German guidelines for operational water-level monitoring (LAWA, 2018), between 57.4 % (ELB) and 89.1 % (LAU) of the measurements fall within ±2 cm of the reference, and 75.3 % of the measurements at GRO satisfy this criterion. The 95th percentile of the absolute differences ranges from 3.0 cm at LAU to 7.0 cm at ELB, showing that large discrepancies are infrequent and that, at the best-performing sites, most errors are limited to a few centimetres.

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f04

Figure 4Multi-site water level time series validation. Absolute error time series (top panels) and daily mean water level time series (bottom panels) comparing wl_opt and wl_ref at ELB, LAU and GRO sites. Shaded areas indicate periods where wl_opt exceeds wl_ref (blue) and where wl_ref exceeds wl_opt (orange).

Download

The error statistics reported in Table 4 are consistent with the patterns shown in Fig. 4, which presents time series comparisons of wl_opt and wl_ref at the ELB, LAU, and GRO monitoring sites for the complete observation periods. For each site, the upper subplot displays the absolute differences as time series. The main subplot presents the time series of daily averaged water levels from both measurements, showing that wl_opt closely tracks wl_ref across a wide range of conditions, including both low-flow and high-flow events, with only centimetric deviations overall. While the largest departures from the reference typically occur during rapid level rises – such as the January 2024 events (4 cm at LAU for water levels near 70 cm, 8 cm at ELB above 150 cm, and 7 cm at GRO above 120 cm). These deviations can be attributed to differences between the measurement concepts, as the reference gauges provide 15 min water levels averaged over 5 min sampling intervals, which tend to smooth rapid fluctuations typical of flood events, whereas the camera-based estimates are derived from instantaneous images.

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f05

Figure 5Validation metrics for LAU monitoring site. (a) Parity plots for all measurements (top left) and daily means (top right, scatter). (b) Bottom panels show detailed residuals analysis for all measurements: distribution histogram (left) and scatter plot vs. water level (right) for daily mean values.

Download

Figure 5 presents the statistical validation for LAU. The parity plots (Fig. 5a) show a close agreement between wl_opt and wl_ref over the full range of observed water levels (3–80 cm), with Pearson = 0.99 and Spearman = 0.97. To keep the comparison legible given the large sample size (n= 64 388), all individual measurements are represented as a hexbin density map, whereas daily means (n= 927) are plotted as individual points, both clustering around the 1 : 1 line. Nash–Sutcliffe efficiency values are also high at all sites, ranging from 0.94 to 0.99 (Table 4). The residuals analysis (Fig. 5b) indicates that 89 % of measurements fall within ±2 cm and that the 95th percentile is ±3.0 cm. A slight systematic pattern is observed, with residuals tending to increase with water level, particularly between 15–30 cm, which may be related to the presence of a step in the measurement wall at LAU. Summary statistics for the three monitoring sites are reported in Table 4, and the corresponding parity and residual plots for ELB and GRO are provided in Appendix C (Figs. C1 and C2, respectively).

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f06

Figure 6Error metrics for LAU monitoring site. (a) MAE distribution (top left) and MAE vs. water level (top right) for all measurements and daily means. (b) sMAPE vs. water level (bottom) for all measurements (left) and daily means (right).

Download

Figure 6a summarises the distribution of absolute errors at LAU, both for all measurements and for daily means, including the 95th percentile and the fraction of values within ±2 cm; equivalent MAE plots for ELB and GRO are provided in Appendix C (Figs. C3a and C4a). When the data are aggregated to daily means, the temporal mismatch between instantaneous wl_opt and averaged wl_ref is partly smoothed and the errors decrease further. Daily MAE is 0.94 cm at LAU and remains below 2.25 and 1.56 cm at ELB and GRO, respectively, while the 95th percentile of the daily absolute differences ranges from 2.21 cm at LAU to 5.62 cm at ELB. At LAU, in particular, 95 % of the daily mean values differ from the gauge record by less than about 2.2 cm and 91.9 % of the days remain within ±2 cm of the reference, indicating that most daily water levels derived from the optical method deviate from the gauge by only a few centimetres. Figure 6b shows the corresponding symmetric mean absolute percentage errors (sMAPE) at LAU for all measurements and for daily means, while the analogous plots for ELB and GRO are again presented in Appendix C (Figs. C3b and C4b). For daily averaged water levels, sMAPE values remain around 5 % at ELB and GRO, whereas values close to 9 % at LAU mainly reflect centimetric fluctuations over very low water levels, where small absolute differences translate into comparatively large relative errors. This behaviour is consistent with the sMAPE–wl_ref relationships, which show decreasing relative errors as water levels increase at the three monitoring sites.

Table 5Statistical comparison between wl_opt and wl_ref for the first 180 d of operation at each station, for individual measurements and daily averaged values.

Download Print Version | Download XLSX

To assess temporal stability and the need for maintenance, the same statistics were computed for the first 180 d of operation at each station (Table 5). For ELB and GRO, all key metrics improve in this initial period compared to the full time series, with lower MAE and 95th percentiles and a higher fraction of daily values within ±2 cm of the reference. At LAU, the performance during the first 180 d is very similar to that of the complete period: individual MAE and the 95th percentile remain essentially unchanged, and daily MAE and the fraction within ±2 cm are only slightly less favourable, indicating that the system maintains a stable accuracy over time.

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f07

Figure 7Automated water-level measurement at peak-flow conditions at LAU (December 2023). The image shows the detected ground control points (crosses), the segmented water-surface boundary (blue line), the upper contour segment used for water-level estimation (red line), and the section of the 3D model colored according to the Z coordinate.

Download

Figure 7 shows the measurement computed at the time of the maximum water level at the LAU station. The figure includes the original image captured by the cameras, the re-projection of the 3D model used to calculate the Z coordinate, the automatically detected GCPs (crosses), the segmented water boundary (blue line), and the specific water-surface segment used to extract the Z coordinate (red points), together with the comparison between wl_opt and wl_ref at the stations where a gauge is available (ELB, LAU and GRO). A video animation illustrating the water-level measurements obtained at the LAU gauging station between 24 and 31 December 2023 is provided in Blanch et al. (2025b) and referenced in the Video Supplement section. Similar peak-flow visualisations for the ELB and GRO stations are presented in Figs. C5 and C6, respectively, in the Appendix.

4 Discussion

The main contribution of this work is the demonstration of a robust solution for water level measurement using low-cost and contactless methods. The systems were tested over 2.5 years at four study sites in Saxony, Germany. The addition of IR lamps has addressed one of the primary limitations of visual methods, enabling reliable measurements at night and allowing for continuous 24 h monitoring. By integrating AI into the photogrammetric workflow, the system proves to be highly robust, functioning effectively also under adverse weather conditions (Fig. 8). The utilization rate of images is very high (average of 98 %), indicating that the system rarely fails to resolve a valid water level. Failures typically occur during periods of extremely poor visibility (e.g., fog, heavy snowfall), when the image-based method is not viable due to inadequate observation of the water surface (Fig. 9a and b).

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f08

Figure 8Successful water level measurements at LAU during snowfall events of varying intensity, showing robust AI segmentation and GCP identification under adverse weather conditions.

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f09

Figure 9Main limitations of the automated camera gauge system. (a, b) LAU during intense snowfall: (a) erroneous AI segmentation due to poor visibility, and (b) same image that despite segmentation errors, a valid water level measurement was obtained. (c) ELB with excessive vegetation covering both riverbanks, obscuring most ground control points and preventing clear water-slope contact delineation.

Download

The success in obtaining water level values under adverse weather conditions is due to our iterative training process of the neural network for water segmentation in images. We developed a model that can handle a wide variety of situations effectively including images that the AI initially struggled to resolve in successive training sessions. It should be noted that KIWA images constitute less than 11 % of the training dataset, suggesting the model's high transferability. For future installations with different cameras and environments, only a small number of manually labelled images will be needed to adapt the model to the new site, because the bulk of the training data come from the RIWA dataset, which spans highly diverse aquatic domains (rivers, lakes, seas, urban environments, etc.). Adding a small percentage of site-specific images ensures maximum segmentation accuracy and, crucially, enables evaluation on the actual target domain, yielding a model tailored to the study area rather than a generalised classifier trained on unrelated scenes (e.g. coastal imagery). Furthermore, modern fine-tuning techniques that retrain only the final network layers can reduce deployment effort even further, potentially avoiding full retraining altogether. This transferability has been successfully demonstrated in practice by Krüger et al. (2024), where the workflow was retrained and applied in a very different environment in Oman, yet still delivered robust results.

The segmentation of images for precise water level measurement presents a significant challenge (i.e., Moghimi et al., 2024; Wagner et al., 2023) because it requires precise identification of the interface between the water body and the riverbanks in the image. This boundary is especially difficult to detect accurately in automatic segmentation processes, since on the one hand it is a natural boundary (i.e., waves, water transparency, vegetation), and on the other hand standard metrics such as the Intersection over the Union (IoU) or the Dice coefficient, and therefore the most common models, give priority to the precise segmentation of the whole object rather than to the precise definition of the contour. In a river context, this delineation is especially challenging due to water transparency at the boundaries, which makes clear delimitation difficult even for human observers. Although we did not employ a contour-centred metric such as the Boundary F1 Score (BFScore), testing various neural network models (Wagner et al., 2023) has allowed us to identify the optimum for correct water segmentation, and thus obtain water level measurements.

Training with night time images proves to be more feasible and accurate because the contrast between the water and the background is more pronounced. Additionally, there are no transparency issues due to the high absorption of infrared light by water, which simplifies the segmentation process. The precision obtained during training for the best-performing model (around 99 %) is consistent with the model's performance throughout the entire time series, demonstrating its ability to consistently segment the water bodies in most images. Ensuring that the mask intersecting with the 3D model accurately represents the water boundary at the moment of image acquisition. The results obtained are in line with the ones provided by Wagner et al. (2023) and Moghimi et al. (2024) in their respective research. However, generalising the use of near-infrared imagery to daytime conditions instead of RGB – with the aim of simplifying segmentation – is not fully reliable. Daylight IR-only thresholding is unreliable because solar NIR and sunlight vary with illumination geometry and surface reflections, which can reduce image sharpness and detail, whereas the nighttime approach operates under controlled illumination and yields more stable contrast.

The automatic identification of GCPs in each image prevents the accuracy from deteriorating over the observation period as effects by experienced camera movements, e.g. due to thermal effects on the camera (Elias et al., 2020), are mitigated. The main issue with the use of GCPs has been their durability in the study areas. Although most remained throughout the 2.5-year study, some were damaged or displaced during high-flow periods or shaded temporally due to vegetation cover, indicating that even with an automated camera gauge, maintenance is necessary. Regarding this maintenance, the difference between Table 4 (whole time series) and Table 5 (first 6 months) shows how the error between the optical values and the reference is higher the longer the time window, and although this difference is multifactorial, undoubtedly the deterioration of the GCPs as well as the age of the camera calibration play an important role. Also, maintenance is necessary due to other key factors such as vegetation. Our results indicate that vegetation significantly impacts deviations to the references by either obscuring GCPs or covering the water-shore contact, leading to irregular segmentation. An effective solution involves automatic regeneration of the 3D models updating the model to the real riverbank situation; however, a crucial aspect of ensuring the robustness of the method is to use measurement areas where the water-slope contact remains unobstructed over time. Figure 4 illustrates how, particularly during summer, sudden changes lead to the system underestimating or overestimating values (notably in ELB, July 2023). These anomalies are related to vegetation changes, with abrupt shifts corresponding to days when landscaping work was performed. Conversely, stations like LAU show more stable values and lower absolute error throughout the year, as the water-shore contact occurs on a concrete wall, providing greater stability over time.

In addition to the accurate segmentation of water and the identification of GCPs, another crucial element for ensuring good water level measurement is the proper calibration of the fixed cameras. This was challenging if the cameras were already installed and operational. In our case, we observe that the ELB camera, which was already in operation and calibrated approximately (i.e., adding the image into the SfM bundle adjustment), shows slightly different precision compared to LAU, where a much more precise calibration was performed (i.e., field board calibration). GRO also underwent field calibration, but the use of a very short focal length and the large distance of the ROI to the camera adversely affects precision. During the two years, the camera calibration files had not been modified. However, assuming a temporal stable interior camera geometry is unlikely amongst other due to the influence of temperature changes (Elias et al., 2020). In the future, refining the results will require the ability to self-calibrate the camera (i.e., obtaining internal parameters) on-site during the analysis of each image (e.g., as done and assessed in Elias et al., 2023).

Table 3 illustrates how cameras with shorter focal lengths exhibit greater re-projection errors. This is expected, as they tend to have higher distortions, particularly at the edges of the image where the GCPs are located. Nevertheless, the results demonstrate that the photogrammetric approach and calibration methods worked sufficiently. Notably, the case of NEU is significant, where despite not having undergone field calibration, the re-projection values are comparable to those obtained at LAU. The acquisition of GCP coordinates and camera calibration are crucial for achieving accurate exterior camera parameter estimation. Errors at these steps affect the accuracy of the reprojection of the 3D model into the 2D space and the subsequent extraction of the Z coordinates.

The creation of the 3D model using SfM algorithms proved to be effective, particularly during normal and high water levels. In urban areas (e.g. GRO and NEU) the models are created using terrestrial cameras only, while at other study sites UAV images are combined. In both cases, the riverbed has to be corrected for refraction influences. In GRO, a larger offset (i.e., 2.5 cm) has to be applied to the wl_opt to minimise the discrepancy between the references zero heights due to the difficulties in accurately modelling the riverbed due to significantly higher water levels on the day of image acquisition. For deep rivers, where SfM cannot be applied because the riverbed is obscured, GNSS support for tracing cross-sections is necessary. In general, a change in absolute error has been observed when the calculated water level intersects the area of the model that is underwater, as the reconstruction quality of this zone is lower. Although this is not a critical issue, if the camera system is used for measuring higher water levels (e.g. floods). Here, we suggest that the 3D model should be generated during periods of lower water levels to improve the accuracies.

The intersection of the water body boundaries with the 3D model determines the calculated water level, making precision in both elements crucial. The standard deviation of the extracted Z coordinates provides an indication of the quality of the intersection, as we assume that the intersection should occur at a constant elevation (i.e., the same Z coordinate). Measurements with high standard deviation suggest difficulty in estimating a stable Z value, which typically indicates either an erroneous segmentation of the water pixels or that the water body-shore contact is not clearly visible in the image (e.g., vegetation obscuring the contact, Fig. 9c). However, the percentage of results filtered by standard deviation (Table 2) is not high, demonstrating the robustness of the method. The application of the Tukey filter based on official reference values has minimal impact on the total number of images used and is only applied to remove erroneous values that may have passed through the standard deviation filter. Typically, these errors are sporadic and also relate to poor segmentation (i.e., segmentation is performed in tiles that may produce horizontal lines intersecting constant Z values), poor image visibility (e.g., presence of ice or animals on the camera), or a water level too low leading to intersections with the 3D model in areas where reconstruction was not optimal.

Regarding the results obtained, the three stations show distinct performance patterns that can be directly related to their installation conditions. At LAU, the precision is consistently high over more than two years, with a daily mean MAE of 0.94 cm and an individual-measurement MAE of 1.26 cm (Table 4). The distribution of individual errors is also compact, with a 95th percentile of 3.0 cm, indicating that most image-based water levels remain very close to the reference even in the upper tail of the error distribution. This behaviour confirms the configuration at LAU as a best-case scenario for future installations: the camera was calibrated on-site, a moderately large focal length was selected to reduce distortions, and the water–shore contact occurs on a concrete wall without vegetation at a moderate distance from the camera, which simplifies segmentation and stabilises the geometry. In contrast, the ELB study site exhibits higher and more variable errors. Over 917 d of monitoring, the daily mean MAE is 2.25 cm and the individual-measurement MAE reaches 2.66 cm, while the 95th percentile of individual errors rises to 7.0 cm (Table 4). These larger deviations are consistent with the challenging water–shore interface at ELB: the contact zone is frequently obscured by vegetation, and the segmentation of the waterline is therefore less stable. Figure 4 – ELB, clearly shows abrupt changes in wl_opt, especially during summer, which coincide with vegetation maintenance and highlight the influence of local riverbank dynamics. Similar vegetation-related issues have been reported in previous image-based water level studies by Eltner et al. (2018) and Peña-Haro et al. (2021). In addition, the presence of a small hydroelectric power plant 100 m upstream introduces rapid changes in water level on the order of decimetres within a few minutes, while the official gauge provides 15 min averages. During such events, the averaged reference value may not represent the instantaneous water level seen by the camera, adding background noise to the error statistics. GRO occupies an intermediate position between LAU and ELB. Here, the daily mean MAE is 1.56 cm and the individual-measurement MAE is 1.75 cm, and the 95th percentile of individual errors reaches 5.0 cm (Table 4). Although the camera was calibrated in the field, the width of the river, the 20 m distance between the camera and the waterline, the use of a wide-angle lens and the partial loss of GCPs increase the sensitivity of the 3D reconstruction and camera orientation to small geometric errors. In addition, the quality of the water–shore contact varies with water level: at higher stages the contact occurs against vertical or near-vertical walls that are easy to delineate, whereas at lower stages it shifts onto a sloping bank with rocks and vegetation, making segmentation more difficult in a way that is similar to ELB. These factors explain why GRO performs clearly better than ELB but does not reach the stability observed at LAU.

During the first 180 d of operation (Table 5), the three stations exhibit very similar behaviour: individual-measurement MAE values range between 1.24 and 1.78 cm, and the 95th percentiles vary between 3.0 and 5.0 cm. For daily means, MAE values are around 0.8–1.6 cm over this initial period. This phase can be regarded as a benchmark for a newly implemented low-cost camera system under favourable and well-maintained conditions, when GCPs are intact, vegetation has limited impact and the camera calibration is recent. When extending the analysis to the full observation period, the degradation of performance at ELB and, to a lesser extent, at GRO becomes apparent (Table 4). The increase in MAE and in the 95th percentiles is consistent with progressive GCP damage or loss, vegetation growth in the water–shore contact zone and ageing camera calibration. In LAU, the daily mean MAE over the whole period is slightly lower than during the first 180 d because the offset used to minimise the differences with the reference is estimated over the entire time series, which partially compensates for initial biases. However, the individual-measurement statistics at LAU still reflect the long-term stability of this installation. Overall, these patterns indicate that camera-based gauges can maintain high accuracy over several years, but only if supported by periodic maintenance and occasional updates of the 3D model and camera calibration.

When comparing our results to previous work conducted at the same study sites, several key differences emerge in terms of precision, operational robustness, and temporal coverage. Eltner et al. (2021) reported an error of 1.1 ± 3.1 cm at ELB using smoothing algorithms, with a best Spearman coefficient of 0.93, but noted significant seasonal variability that affected measurement consistency. Zamboni et al. (2025) proposed a computationally efficient method avoiding model training, achieving mean absolute errors of 2.1 cm (LAU) and 2.9 cm (ELB) with RMSE values ranging from 1.6 to 3.2 cm across their three camera-gauge stations (Wesenitz, Elbersdorf, Lauenstein). While both studies demonstrate valuable methodological contributions, neither addressed night-time observations, long-term resilience to adverse weather conditions, or extended time-series stability. Our approach builds upon this foundation by maintaining comparable or improved precision while systematically evaluating performance over multi-year deployments under full operational conditions at these same locations, including challenging scenarios that previous work did not consider.

Compared to other image-based contactless systems deployed at different sites, our results achieve precision levels competitive with existing methods while extending the evaluation to multi-year continuous operation under full 24/7 conditions, including nighttime and adverse weather. Short-term experimental studies typically cover periods from hours to weeks: Erfani et al. (2023) obtained RMSE values between 1.5 and 2.9 cm over a few hours at Rocky Branch (South Carolina), Muhadi et al. (2021) achieved Spearman correlations of 0.91 using DeepLabv3 in Malaysia, Wang et al. (2024) obtained 0.87–0.94 over two weeks in UK rivers, and Moy de Vitry et al. (2019) reached 0.75–0.85 with U-Net on Swiss CCTV footage using the SOFI trend index. Among longer-term deployments, Vandaele et al. (2021) reported Pearson correlations of 0.94–0.96 over one year using landmark-based water level estimation on UK rivers, while LSPIV-based systems have demonstrated extended capabilities: Ran et al. (2016) obtained RMSE of 3.7 cm over flood events in a Chinese mountainous stream, Stumpf et al. (2016) achieved MAPE of 9 % over 2.5 years analyzing more than 4000 videos in a tropical mountain river (Réunion Island), and Peña-Haro et al. (2021) reported RMSE of 6.8 cm across multiple Swiss Alpine sites with their operational DischargeKeeper system. Our work bridges the gap between experimental validation and operational deployment, achieving Spearman correlations of 0.94–0.97 and Pearson correlations of 0.97–0.99 over multi-year continuous monitoring that consistently captures different hydrological regimes and challenging environmental conditions – a combination of accuracy, duration, and robustness not previously demonstrated in camera-based systems.

The results of this study are consistent with the German manual for river water level measurements, requiring water levels with accuracies below 2.0 cm as an acceptable systematic error (LAWA, 2018). When considering the entire observation period (Table 4), both LAU and GRO still maintain a high proportion of daily means within ±2 cm of the reference (91.9 % and 71.0 %, respectively), confirming that the camera-based gauges can meet hydrometric standards over multi-year deployments. In contrast, ELB clearly performs worse in this respect, with only 56.9 % of daily means within the 2 cm range, which is consistent with the site-specific issues at this station already discussed. Nevertheless, during the first 6 months of operation all three stations showed substantially higher fractions of individual measurements within ±2 cm (Table 5), indicating that under favourable and well-maintained conditions the system can achieve very similar performance across different sites. However, the error levels observed here are consistently lower than those relevant for flood detection and AI-based observation, making our approach suitable and advantageous for continuous river monitoring and early warning systems. In addition, the stability achieved in the water level estimates provides a solid basis for further developments in river monitoring through camera-based gauges and video series, enabling future integration of image-based methods to derive flow and velocity.

5 Conclusion

This study demonstrates the effectiveness of an AI-enhanced, image-based camera gauge for long-term, near real-time river water level monitoring. Over a 2.5-year period, the system delivered accurate water level estimates at three contrasting sites, with daily mean MAE values as low as 0.94 cm and individual-measurement MAE values between 1.26 and 2.66 cm, which are consistent with, and in most cases comfortably below, the 2.0 cm accuracy requirement of the German manual for river water level measurements. The combination of neural networks for water segmentation and GCP identification with a photogrammetric workflow enabled fully automatic processing of large image datasets under a wide range of hydrological and environmental conditions. The approach relies on low-cost hardware and standard surveillance cameras, which facilitates deployment at additional sites and integration into existing hydrometric networks. The installation of IR lamps, combined with the surveillance cameras' ability to capture nighttime images, mitigated one of the major limitations of image-based methods, namely their restriction to daylight, and enabled continuous 24/7 water level monitoring, which is particularly relevant for flood early-warning applications. The observed increase in inaccuracies over time highlights the need for proper maintenance of both the environment (e.g. vegetation) and the cameras (e.g. calibration). Future work aimed at minimising these maintenance requirements, for example by regenerating 3D models to adapt to terrain changes and by implementing automatic camera calibration procedures, is expected to further improve measurement quality and robustness.

Appendix A

A1 Nomenclature and Abbreviations

ADE20K Semantic scene dataset used for segmentation benchmarks
AI Artificial Intelligence
BFScore Boundary F1 Score
CNN Convolutional Neural Network
DSLR Digital Single-Lens Reflex
GCP Ground Control Point
GNSS Global Navigation Satellite System
IoU Intersection over Union
IQR Interquartile Range (Q3–Q1)
IR Infrared
KNNImputer k-Nearest Neighbours Imputer algorithm (scikit-learn)
LBWLE Landmark-Based Water Level Estimation
MAE Mean Absolute Error
MVS Multi-View Stereo
NSE Nash–Sutcliffe Efficiency
OCR Optical Character Recognition
R-CNN Region-based Convolutional Neural Network
ResNeXt50 Residual network with grouped convolutions (50 layers)
ResUnet Residual U-Net neural network
RGB Red, Green, Blue
RIWA River Water Segmentation Dataset
RMSE Root Mean Square Error
ROI Region of Interest
RTK-GNSS Real-Time Kinematic Global Navigation Satellite System
SAM Segment Anything Model
SfM Structure from Motion
SIFT Scale-Invariant Feature Transform
sMAPE Symmetric Mean Absolute Percentage Error
SOFI Static Observer Flooding Index
STCN Space-Time Correspondence Networks
SURF Speeded Up Robust Features
PyBathySfM Python Bathymetric Structure-from-Motion software package
UAV Unmanned Aerial Vehicle
U-Net Convolutional neural network architecture for image segmentation
UPerNet Unified Perceptual Parsing Network
WaterNet Dataset for water segmentation

A2 Study-Specific Abbreviations

KIWA Artificial Intelligence for Flood Warning project (funding/project framework of this study)
ELB Elbersdorf gauging station at the Wesenitz River
GRO Großschönau 2 gauging station at the Mandau River
LAU Lauenstein 4 gauging station at the Müglitz River
NEU Neukirch site at the Wesenitz River
h_opt Water surface elevation measured by the optical system (in metres above reference)
h_opt0 Reference zero height in the riverbed of the 3D model (optical system)
h_ref0 Reference zero height of the official gauging station
wl_opt Water level calculated by the optical system (in centimetres)
wl_ref Water level from the official reference gauge (in centimetres)
Appendix B: Flowchart plot
https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f10

Figure B1Flowchart summarising the complete process from image acquisition to water level determination. The chart outlines the steps from the initial JPG image captured by the camera to the final extraction of water level data.

Download

Appendix C: ELB and GRO Results
https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f11

Figure C1Validation metrics for ELB monitoring site. (a) Parity plots for all measurements (top left) and daily means (top right, scatter). (b) Bottom panels show detailed residuals analysis for all measurements: distribution histogram (left) and scatter plot vs. water level (right) for daily mean values.

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f12

Figure C2Validation metrics for GRO monitoring site. (a) Parity plots for all measurements (top left) and daily means (top right, scatter). (b) Bottom panels show detailed residuals analysis for all measurements: distribution histogram (left) and scatter plot vs. water level (right) for daily mean values.

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f13

Figure C3Error metrics for ELB monitoring site. (a) MAE distribution (top left) and MAE vs. water level (top right) for all measurements and daily means. (b) sMAPE vs. water level (bottom) for all measurements (left) and daily means (right).

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f14

Figure C4Error metrics for GRO monitoring site. (a) MAE distribution (top left) and MAE vs. water level (top right) for all measurements and daily means. (b) sMAPE vs. water level (bottom) for all measurements (left) and daily means (right).

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f15

Figure C5Automated water level measurement at peak flow conditions (ELB, December 2023). The image shows the automated identification of ground control points (GCPs, crosses), the segmented water body boundary (blue line), and the upper contour segment used for water level calculation (red line).

Download

https://hess.copernicus.org/articles/30/797/2026/hess-30-797-2026-f16

Figure C6Automated water level measurement at peak flow conditions (GRO, December 2023). The image shows the automated identification of ground control points (GCPs, crosses), the segmented water body boundary (blue line), and the upper contour segment used for water level calculation (red line).

Download

Code and data availability

The source code used in this study is publicly accessible in a Zenodo repository (Blanch, 2025): https://doi.org/10.5281/zenodo.17675672. The RIWA dataset used for the preliminary training is available in https://doi.org/10.34740/kaggle/dsv/4901781 (Blanch et al., 2023b, login required).

Video supplement

The video supplement presents an animation of the water level measurements obtained at the LAU gauging station from 24 to 31 December 2023, using the methodology described in this publication. It is accessible in the Zenodo repository: https://doi.org/10.5281/zenodo.14875801 (Blanch et al., 2025b).

Author contributions

All authors contributed greatly to the work. XB: conceptualization, methodology, software, investigation, writing (original draft), figures, data acquisition; AE: conceptualization, methodology, writing (review and editing), supervision, funding acquisition; JG: conceptualization, supervision, writing (review), data acquisition, funding acquisition; RH: data acquisition, writing (review), supervision. All authors have read and agreed to the published version of the paper.

Competing interests

The contact author has declared that none of the authors has any competing interests.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. The authors bear the ultimate responsibility for providing appropriate place names. Views expressed in the text are those of the authors and do not necessarily reflect the views of the publisher.

Acknowledgements

The authors would like to thank the Saxon state company for the environment and agriculture (Staatliche Betriebsgesellschaft für Umwelt und Landwirtschaft Sachsen), Germany for their support and cooperation in this study.

Financial support

This research has been supported by the Bundesministerium für Forschung, Technologie und Raumfahrt (grant nos. 13N15542 and 13N15543) under the announcement “Artificial Intelligence in Civil Security Research”, which is part of the Federal Government's “Research for Civil Security” programme.

Review statement

This paper was edited by Ralf Loritz and reviewed by Salvatore Manfreda, Riccardo Taormina, and one anonymous referee.

References

Bartos, M., Wong, B., and Kerkez, B.: Open storm: a complete framework for sensing and control of urban watersheds, Environ. Sci. Water Res. Technol., 4, 346–358, https://doi.org/10.1039/C7EW00374A, 2018. 

Blanch, X., Guinau, M., Eltner, A., and Abellan, A.: Fixed photogrammetric systems for natural hazard monitoring with high spatio-temporal resolution, Nat. Hazards Earth Syst. Sci., 23, 3285–3303, https://doi.org/10.5194/nhess-23-3285-2023, 2023a. 

Blanch, X., Wagner, F., and Eltner, A.: River Water Segmentation Dataset (RIWA), Kaggle [data set], https://doi.org/10.34740/kaggle/dsv/4901781, 2023b. 

Blanch, X., Guinau, M., Eltner, A., and Abellan, A.: A cost-effective image-based system for 3D geomorphic monitoring: An application to rockfalls, Geomorphology, 449, 109065, https://doi.org/10.1016/j.geomorph.2024.109065, 2024. 

Blanch, X., Jäschke, A., Elias, M., and Eltner, A.: Subpixel Automatic Detection of GCP Coordinates in Time-Lapse Images Using a Deep Learning Keypoint Network, IEEE Trans. Geosci. Remote Sens., 63, 1–14, https://doi.org/10.1109/TGRS.2024.3514854, 2025a. 

Blanch, X., Eltner, A., Grundmann, J., and Hedel, R.: Water level results at Lauenstein gauge station – KIWA Project, Zenodo [video], https://doi.org/10.5281/zenodo.14875801, 2025b. 

Blanch, X.: KIWA Software for Water Level (wl_opt) – KIWA Project, in: Hydrology and Earth System Sciences (HESS), Zenodo [code], https://doi.org/10.5281/zenodo.17675672, 2025. 

Buslaev, A., Parinov, A., Khvedchenya, E., Iglovikov, V. I., and Kalinin, A. A.: Albumentations: fast and flexible image augmentations, Information, 11, 125, https://doi.org/10.3390/info11020125, 2020. 

Chureesampant, K. and Susaki, J.: Automatic GCP Extraction of Fully Polarimetric SAR Images, IEEE Trans. Geosci. Remote Sens., 52, 137–148, https://doi.org/10.1109/TGRS.2012.2236890, 2014. 

Dietrich, J. T.: Bathymetric Structure-from-Motion: extracting shallow stream bathymetry from multi-view stereo photogrammetry, Earth Surf. Process. Landf., 42, 355–364, https://doi.org/10.1002/esp.4060, 2017. 

Elias, M., Kehl, C., and Schneider, D.: Photogrammetric water level determination using smartphone technology, Photogramm. Rec., 34, 198–223, https://doi.org/10.1111/phor.12280, 2019. 

Elias, M., Eltner, A., Liebold, F., and Maas, H.-G.: Assessing the Influence of Temperature Changes on the Geometric Stability of Smartphone- and Raspberry Pi Cameras, Sensors, 20, 643, https://doi.org/10.3390/s20030643, 2020. 

Elias, M., Weitkamp, A., and Eltner, A.: Multi-modal image matching to colorize a SLAM based point cloud with arbitrary data from a thermal camera, ISPRS Open J. Photogramm. Remote Sens., 9, 100041, https://doi.org/10.1016/j.ophoto.2023.100041, 2023. 

Eltner, A. and Sofia, G.: Chapter 1 – Structure from motion photogrammetric technique, in: Developments in Earth Surface Processes, vol. 23, edited by: Tarolli, P. and Mudd, S. M., Elsevier, 1–24, https://doi.org/10.1016/B978-0-444-64177-9.00001-1, 2020. 

Eltner, A., Kaiser, A., Abellan, A., and Schindewolf, M.: Time lapse structure-from-motion photogrammetry for continuous geomorphic monitoring, Earth Surf. Process. Landf., 42, 2240–2253, https://doi.org/10.1002/esp.4178, 2017. 

Eltner, A., Elias, M., Sardemann, H., and Spieler, D.: Automatic Image-Based Water Stage Measurement for Long-Term Observations in Ungauged Catchments, Water Resour. Res., 54, https://doi.org/10.1029/2018WR023913, 2018. 

Eltner, A., Sardemann, H., and Grundmann, J.: Technical Note: Flow velocity and discharge measurement in rivers using terrestrial and unmanned-aerial-vehicle imagery, Hydrol. Earth Syst. Sci., 24, 1429–1445, https://doi.org/10.5194/hess-24-1429-2020, 2020. 

Eltner, A., Bressan, P. O., Akiyama, T., Gonçalves, W. N., and Marcato Junior, J.: Using Deep Learning for Automatic Water Stage Measurements, Water Resour. Res., 57, e2020WR027608, https://doi.org/10.1029/2020WR027608, 2021. 

Epple, L., Grothum, O., Bienert, A., and Eltner, A.: Decoding rainfall effects on soil surface changes: Empirical separation of sediment yield in time-lapse SfM photogrammetry measurements, Soil Tillage Res., 248, 106384, https://doi.org/10.1016/j.still.2024.106384, 2025. 

Erfani, S. M. H., Smith, C., Wu, Z., Shamsabadi, E. A., Khatami, F., Downey, A. R. J., Imran, J., and Goharian, E.: Eye of Horus: a vision-based framework for real-time water level measurement, Hydrol. Earth Syst. Sci., 27, 4135–4149, https://doi.org/10.5194/hess-27-4135-2023, 2023. 

Fernandes, F. E., Nonato, L. G., and Ueyama, J.: A river flooding detection system based on deep learning and computer vision, Multimed. Tools Appl., 81, 40231–40251, https://doi.org/10.1007/s11042-022-12813-3, 2022. 

Grundmann, J., Blanch, X., Kutscher, A., Hedel, R., and Eltner, A.: Towards a comprehensive optical workflow for monitoring and estimation of water levels and discharge in watercourses, EGU General Assembly 2024, Vienna, Austria, 14–19 April 2024, EGU24-12507, https://doi.org/10.5194/egusphere-egu24-12507, 2024. 

Herschy, R. W.: Streamflow Measurement, 3rd edn., CRC Press, London, 536 pp., https://doi.org/10.1201/9781482265880, 2008. 

Iglhaut, J., Cabo, C., Puliti, S., Piermattei, L., O'Connor, J., and Rosette, J.: Structure from Motion Photogrammetry in Forestry: a Review, Curr. For. Rep., 5, 155–168, https://doi.org/10.1007/s40725-019-00094-3, 2019. 

Ioli, F., Bruno, E., Calzolari, D., Galbiati, M., Mannocchi, A., Manzoni, P., Martini, M., Bianchi, A., Cina, A., De Michele, C., and Pinto, L.: A Replicable Open-Source Multi-Camera System for Low-Cost 4d Glacier Monitoring, Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci., XLVIII-M-1–2023, 137–144, https://doi.org/10.5194/isprs-archives-XLVIII-M-1-2023-137-2023, 2023. 

Isikdogan, F., Bovik, A. C., and Passalacqua, P.: Surface Water Mapping by Deep Learning, IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens., 10, 4909–4918, https://doi.org/10.1109/JSTARS.2017.2735443, 2017. 

ISO 18365: Hydrometry – Selection, establishment and operation of a gauging station, ISO (International Organization for Standardization), Geneva, Switzerland, 2013. 

Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A. C., Lo, W.-Y., Dollár, P., and Girshick, R.: Segment Anything, arXiv [preprint], https://doi.org/10.48550/arXiv.2304.02643, 2023. 

Krüger, R., Blanch, X., Grundmann, J., Al-Rawas, G., and Eltner, A.: Low-Cost sensors for measuring wadi discharge – a Raspberry Pi based seismometer and time-lapse camera setup, Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci., XLVIII-2/W8-2024, 243–250, https://doi.org/10.5194/isprs-archives-XLVIII-2-W8-2024-243-2024, 2024. 

Kuo, L.-C. and Tai, C.-C.: Robust Image-Based Water-Level Estimation Using Single-Camera Monitoring, IEEE Trans. Instrum. Meas., 71, 1–11, https://doi.org/10.1109/TIM.2022.3161691, 2022. 

LAWA – Bund/Länderarbeitsgemeinschaft Wasser: Leitfaden zur Hydrometrie des Bundes und der Länder – Pegelhandbuch, https://www.lawa.de/documents/02_anhang_2_lawa_pegelhandbuch_2_3_1552303807.pdf (last access: 11 February 2026), 2018. 

Leduc, P., Ashmore, P., and Sjogren, D.: Technical note: Stage and water width measurement of a mountain stream using a simple time-lapse camera, Hydrol. Earth Syst. Sci., 22, 1–11, https://doi.org/10.5194/hess-22-1-2018, 2018. 

Li, D., Wang, Z., Chen, Y., Jiang, R., Ding, W., and Okumura, M.: A Survey on Deep Active Learning: Recent Advances and New Frontiers, arXiv [preprint], https://doi.org/10.48550/arXiv.2405.00334, 15 July 2024. 

Liang, Y., Jafari, N., Luo, X., Chen, Q., Cao, Y., and Li, X.: WaterNet: An adaptive matching pipeline for segmenting water with volatile appearance, Comput. Vis. Media, 6, 65–78, https://doi.org/10.1007/s41095-020-0156-x, 2020. 

Liebold, F., Mader, D., Sardemann, H., Eltner, A., and Maas, H.-G.: A Bi-Radial Model for Lens Distortion Correction of Low-Cost UAV Cameras, Remote Sens., 15, 5283, https://doi.org/10.3390/rs15225283, 2023. 

Maalek, R. and Lichti, D. D.: Robust detection of non-overlapping ellipses from points with applications to circular target extraction in images and cylinder detection in point clouds, ISPRS J. Photogramm. Remote Sens., 176, 83–108, https://doi.org/10.1016/j.isprsjprs.2021.04.010, 2021. 

Manfreda, S., Miglino, D., Saddi, K. C., Jomaa, S., Eltner, A., Perks, M., Peña-Haro, S., Bogaard, T., Van Emmerik, T. H. M., Mariani, S., Maddock, I., Tauro, F., Grimaldi, S., Zeng, Y., Gonçalves, G., Strelnikova, D., Bussettini, M., Marchetti, G., Lastoria, B., Su, Z., and Rode, M.: Advancing river monitoring using image-based techniques: challenges and opportunities, Hydrol. Sci. J., 69, 657–677, https://doi.org/10.1080/02626667.2024.2333846, 2024. 

Miglino, D., Jomaa, S., Rode, M., Saddi, K. C., Isgrò, F., and Manfreda, S.: Technical note: Image processing for continuous river turbidity monitoring – full-scale tests and potential applications, Hydrol. Earth Syst. Sci., 29, 4133–4151, https://doi.org/10.5194/hess-29-4133-2025, 2025. 

Moghimi, A., Welzel, M., Celik, T., and Schlurmann, T.: A Comparative Performance Analysis of Popular Deep Learning Models and Segment Anything Model (SAM) for River Water Segmentation in Close-Range Remote Sensing Imagery, IEEE Access, 1–1, https://doi.org/10.1109/ACCESS.2024.3385425, 2024. 

Morgenschweis, G.: Hydrometrie: Theorie und Praxis der Durchflussmessung in offenen Gerinnen, Springer, Berlin, Heidelberg, https://doi.org/10.1007/978-3-662-55314-5, 2018. 

Moy de Vitry, M., Kramer, S., Wegner, J. D., and Leitão, J. P.: Scalable flood level trend monitoring with surveillance cameras using a deep convolutional neural network, Hydrol. Earth Syst. Sci., 23, 4621–4634, https://doi.org/10.5194/hess-23-4621-2019, 2019. 

Muhadi, N. A., Abdullah, A. F., Bejo, S. K., Mahadi, M. R., and Mijic, A.: Deep Learning Semantic Segmentation for Water Level Estimation Using Surveillance Camera, Appl. Sci., 11, 9691, https://doi.org/10.3390/app11209691, 2021. 

Mydlarz, C., Sai Venkat Challagonda, P., Steers, B., Rucker, J., Brain, T., Branco, B., Burnett, H. E., Kaur, A., Fischman, R., Graziano, K., Krueger, K., Hénaff, E., Ignace, V., Jozwiak, E., Palchuri, J., Pierone, P., Rothman, P., Toledo-Crow, R., and Silverman, A. I.: FloodNet: Low-Cost Ultrasonic Sensors for Real-Time Measurement of Hyperlocal, Street-Level Floods in New York City, Water Resour. Res., 60, e2023WR036806, https://doi.org/10.1029/2023WR036806, 2024. 

Pan, J., Yin, Y., Xiong, J., Luo, W., Gui, G., and Sari, H.: Deep Learning-Based Unmanned Surveillance Systems for Observing Water Levels, IEEE Access, 6, 73561–73571, https://doi.org/10.1109/ACCESS.2018.2883702, 2018. 

Peña-Haro, S., Carrel, M., Lüthi, B., Hansen, I., and Lukes, R.: Robust Image-Based Streamflow Measurements for Real-Time Continuous Monitoring, Front. Water, 3, 766918, https://doi.org/10.3389/frwa.2021.766918, 2021. 

Ran, Q., Li, W., Liao, Q., Tang, H., and Wang, M.: Application of an automated LSPIV system in a mountainous stream for continuous flood flow measurements, Hydrol. Process., 30, 3014–3029, https://doi.org/10.1002/hyp.10836, 2016. 

Ridolfi, E. and Manciola, P.: Water Level Measurements from Drones: A Pilot Case Study at a Dam Site, Water, 10, 297, https://doi.org/10.3390/w10030297, 2018. 

Ronneberger, O., Fischer, P., and Brox, T.: U-Net: Convolutional Networks for Biomedical Image Segmentation, in: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, vol. 9351, edited by: Navab, N., Hornegger, J., Wells, W. M., and Frangi, A. F., Springer International Publishing, Cham, 234–241, https://doi.org/10.1007/978-3-319-24574-4_28, 2015. 

Smith, M. W., Carrivick, J. L., and Quincey, D. J.: Structure from motion photogrammetry in physical geography, Prog. Phys. Geogr. Earth Environ., 40, 247–275, https://doi.org/10.1177/0309133315615805, 2016. 

Stumpf, A., Augereau, E., Delacourt, C., and Bonnier, J.: Photogrammetric discharge monitoring of small tropical mountain rivers: A case study at Rivière des Pluies, Réunion Island, Water Resour. Res., 52, 4550–4570, https://doi.org/10.1002/2015WR018292, 2016.  

Vandaele, R., Dance, S. L., and Ojha, V.: Deep learning for automated river-level monitoring through river-camera images: an approach based on water segmentation and transfer learning, Hydrol. Earth Syst. Sci., 25, 4435–4453, https://doi.org/10.5194/hess-25-4435-2021, 2021. 

Wagner, F., Eltner, A., and Maas, H.-G.: River water segmentation in surveillance camera images: A comparative study of offline and online augmentation using 32 CNNs, Int. J. Appl. Earth Obs. Geoinformation, 119, 103305, https://doi.org/10.1016/j.jag.2023.103305, 2023. 

Wang, Z., Lyu, H., Zhou, S., and Zhang, C.: Toward Transferable Water Level Trend Monitoring Using River Cameras: Integrating Domain-Specific Models and Segment Anything Model (SAM) for Water Segmentation, ESS Open Archive, https://doi.org/10.22541/essoar.171104253.38492313/v2, 25 March 2024. 

Westoby, M. J., Brasington, J., Glasser, N. F., Hambrey, M. J., and Reynolds, J. M.: “Structure-from-Motion” photogrammetry: A low-cost, effective tool for geoscience applications, Geomorphology, 179, 300–314, https://doi.org/10.1016/j.geomorph.2012.08.021, 2012. 

Xiao, T., Liu, Y., Zhou, B., Jiang, Y., and Sun, J.: Unified Perceptual Parsing for Scene Understanding, in: Computer Vision – ECCV 2018, vol. 11209, edited by: Ferrari, V., Hebert, M., Sminchisescu, C., and Weiss, Y., Springer International Publishing, Cham, 432–448, https://doi.org/10.1007/978-3-030-01228-1_26, 2018. 

Young, D. S., Hart, J. K., and Martinez, K.: Image analysis techniques to estimate river discharge using time-lapse cameras in remote locations, Comput. Geosci., 76, 1–10, https://doi.org/10.1016/j.cageo.2014.11.008, 2015. 

Zamboni, P. A. P., Gorriz, X. B., Junior, J. M., Nunes, W., and Eltner, A.: Do we need to label large datasets for river water segmentation? Benchmark and stage estimation with minimum to non-labeled image time series, International Journal of Remote Sensing, 46, 2719–2747, https://doi.org/10.1080/01431161.2025.2457131, 2025. 

Zhou, B., Zhao, H., Puig, X., Xiao, T., Fidler, S., Barriuso, A., and Torralba, A.: Semantic Understanding of Scenes Through the ADE20K Dataset, Int. J. Comput. Vis., 127, 302–321, https://doi.org/10.1007/s11263-018-1140-0, 2019. 

Zhou, N., Chen, H., Chen, N., and Liu, B.: An Approach Based on Video Image Intelligent Recognition for Water Turbidity Monitoring in River, ACS EST Water, 4, 543–554, https://doi.org/10.1021/acsestwater.3c00596, 2024. 

Download
Short summary
This study presents a low-cost, automated system for monitoring river water levels using cameras and AI. By combining AI-based image analysis with photogrammetry, it accurately measures water levels in real-time, even in challenging conditions. Tested over 2.5 years at four sites, it achieved high accuracy (errors of 1.0–2.3 cm) and processed over 219 000 images. Its resilience makes it ideal for flood detection and water management in remote areas.
Share