Water body extraction from remote sensing images in complex backgrounds is crucial for environmental monitoring, disaster management, and urban planning. Although existing water body extraction algorithms for remote sensing images offer robust tools for monitoring, they still face challenges, including the inability to extract fine water bodies and the frequent omission of water body edges in complex backgrounds such as dense vegetation, varying terrain, or cloud interference. In response, we introduce an architecture named WatNet to enhance the precision of water body extraction in complex environments. WatNet comprises three main modules: the global multi-attention fusion module (GMAF), the water forward network module (WFN), and the edge focus attention module (EFA). The GMAF module enhances the model’s capability to capture global information through multi-head self-attention and convolutional attention modules, improving the overall feature extraction of water bodies. The WFN module utilizes depth-wise separable convolution and attention mechanisms to enhance the capture of local features in fine water bodies. The EFA module significantly improves the clarity and accuracy of water body boundaries through refined edge detection. Experiments on the LoveDA Remote Sensing Land Cover (LoveDA), Qinghai–Tibet Plateau (QTPL), and Wuhan dense labeling (WHDLD) datasets show that WatNet outperforms the mainstream methods in precision, recall, overall accuracy (OA), F1 score, and mean Intersection over Union (mIoU). On the LoveDA dataset, WatNet’s mIoU improved by 1.24% compared with the second-best method, by 0.4% on the QTPL dataset, and by 0.51% on the WHDLD dataset. These results validate the WatNet’s effectiveness and robustness in water body extraction tasks under various environmental conditions.
When the Visual-Inertial Odometry (VIO) is started, its Inertial Measurement Unit (IMU) lacks acceleration incentive, which will result in poor orientation estimation accuracy during initialization, or even initialization failure. Therefore, a visual priori map-assisted monocular location algorithm based on 3D spatial straight lines is proposed. Firstly, the monocular image data of the surrounding environment were extracted through the Line Segment Detection algorithm (LSD), and high precision 2D line features were selected according to the length of the line and the number of surrounding point features. The 3D spatial lines of the surrounding environment were obtained using the line and surface intersection method. Construct a visual prior map with 3D spatial straight lines. Secondly, the constructed visual prior map is used as the online monocular VIO pose estimation for the global map. Based on the straight-line feature matching algorithm and the 3D space straight line depth information as additional constraints, the 2D straight-line feature in the monocular VIO's current field of vision is matched with the 3D space straight line in the visual prior map. The matching results were used as global constraints to optimize the monocular VIO pose. Tests on EUROC and TUM common data sets show that the 3D spatial straight line based visual prior map can effectively correct the pose during the monocular VIO initialization stage. Compared with the VINS-Mono localization algorithm, this algorithm can effectively improve the pose estimation accuracy during VIO initialization and reduce the overall trajectory positioning error.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.