• 中国科技核心期刊
  • JST收录期刊
  • Scopus收录期刊
  • DOAJ收录期刊
Turn off MathJax
Article Contents
BAI Zhongyu, XU Hongli, RU Jingyu, QIU Shaoxiong. Style Transfer-Based Data Augmentation for Side-Scan Sonar Image Classification[J]. Journal of Unmanned Undersea Systems. doi: 10.11993/j.issn.2096-3920.2025-0045
Citation: BAI Zhongyu, XU Hongli, RU Jingyu, QIU Shaoxiong. Style Transfer-Based Data Augmentation for Side-Scan Sonar Image Classification[J]. Journal of Unmanned Undersea Systems. doi: 10.11993/j.issn.2096-3920.2025-0045

Style Transfer-Based Data Augmentation for Side-Scan Sonar Image Classification

doi: 10.11993/j.issn.2096-3920.2025-0045
  • Received Date: 2025-03-13
  • Accepted Date: 2025-04-16
  • Rev Recd Date: 2025-04-07
  • Available Online: 2025-07-07
  • Side-scan sonar (SSS) has been widely employed in ocean exploration due to its stability and efficiency on autonomous underwater vehicles (AUVs). However, the difficulty in acquiring SSS images and the limited availability of training samples significantly limit the performance of the deep neural network (DNN)-based SSS image classification. To address this challenge, this paper proposes a multi-scale attention network (MSANet) that utilizes optical-acoustic image pairs for data augmentation to enhance the generalization ability of SSS image classification models. Specifically, shallow and deep features are extracted from different layers of the encoder to comprehensively capture content and style information. Subsequently, a multi-scale attention module (MSAM) is introduced to extract both local and global contextual information of style features along the channel dimension. These style features are then effectively fused with optical features to achieve precise spatial alignment of optical and acoustic features. Finally, the fused features from different layers are rescaled and fed into a decoder to generate high-fidelity synthetic SSS image samples, which are used to train the SSS image classification network. Experimental results on real-world SSS datasets demonstrate that the proposed style transfer approach effectively generates high-quality synthetic SSS image samples, thereby improving the DNN-based SSS image classification performance.

     

  • loading
  • [1]
    郝紫霄, 王琦. 基于声呐图像的水下目标检测研究综述[J]. 水下无人系统学报, 2023, 31(2): 339-348.

    HAO Z X, WANG Q. Underwater target detection based on sonar image[J]. Journal of Unmanned Undersea Systems, 2023, 31(2): 339-348.
    [2]
    朱兆彤, 付学志, 胡友峰. 一种利用迁移学习训练卷积神经网络的声呐图像识别方法[J]. 水下无人系统学报, 2020, 28(1): 89-96.

    ZHU Z T, FU X Z, HU Y F. A sonar image recognition method based on convolutional neural network trained through transfer learning[J]. Journal of Unmanned Undersea Systems, 2020, 28(1): 89-96.
    [3]
    ZHU B, WANG X, CHU Z, et al. Active learning for recognition of shipwreck target in side-scan sonar image[J]. Remote Sensing, 2019, 11(3): 243. doi: 10.3390/rs11030243
    [4]
    LI C, YE X, CAO D, et al. Zero shot objects classification method of side scan sonar image based on synthesis of pseudo samples[J]. Applied Acoustics, 2021, 173: 107691. doi: 10.1016/j.apacoust.2020.107691
    [5]
    NAYAK N, NARA M, GAMBIN T, et al. Machine learning techniques for AUV side-scan sonar data feature extraction as applied to intelligent search for underwater archaeological sites[C]//Field and Service Robotics: Results of the 12th International Conference. Singapore: Springer, 2021: 219-233.
    [6]
    JIANG Y, KU B, KIM W, et al. Side-scan sonar image synthesis based on generative adversarial network for images in multiple frequencies[J]. IEEE Geoscience and Remote Sensing Letters, 2020, 18(9): 1505-1509.
    [7]
    HUANG C, ZHAO J, YU Y, et al. Comprehensive sample augmentation by fully considering SSS imaging mechanism and environment for shipwreck detection under zero real samples[J]. IEEE Transactions on Geoscience and Remote Sensing, 2021, 60: 1-14.
    [8]
    SIMONYAN K, ZISSERMAN A. Very deep convolutional networks for large-scale image recognition[EB/OL]. [2015-04-10][2025-03-10]. https://arxiv.org/abs/1409.1556.
    [9]
    HUANG X, BELONGIE S. Arbitrary style transfer in real-time with adaptive instance normalization[C]//2017 IEEE International Conference on Computer Vision(ICCV). Venice, Italy: ICCV, 2017: 1510-1519.
    [10]
    DAI Y, GIESEKE F, OEHMCKE S, et al. Attentional feature fusion[C]//2021 IEEE Winter Conference on Applications of Computer Vision(WACV). Waikoloa, USA: WACV, 2021: 3559-3568.
    [11]
    CAO Y, XU J, LIN S, et al. Gcnet: Non-local networks meet squeeze-excitation networks and beyond[C]//2019 IEEE/CVF International Conference on Computer Vision Workshop(ICCVW), Seoul, Korea (South): ICCVW, 2019: 1971-1980.
    [12]
    NOMAN M, STANKOVIC V, TAWFIK A. Object detection techniques: Overview and performance comparison[C]//2019 IEEE International Symposium on Signal Processing and Information Technology(ISSPIT), Ajman, United Arab Emirates: ISSPIT, 2019: 1-5.
    [13]
    PHILLIPS F, MACKINTOSH B. Wiki art gallery, inc.: A case for critical thinking[J]. Issues in Accounting Education, 2011, 26(3): 593-608. doi: 10.2308/iace-50038
    [14]
    LI Y, FANG C, YANG J, et al. Universal style transfer via feature transforms[J]. Advances in neural information processing systems, 2017, 386-396.
    [15]
    HUANG X, BELONGIE S. Arbitrary style transfer in real-time with adaptive instance normalization[C]//2017 IEEE International Conference on Computer Vision(ICCV), Venice, Italy: ICCV, 2017: 1510-1519.
    [16]
    PARK D Y, LEE K H. Arbitrary style transfer with style-attentional networks[C]//2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition(CVPR), Long Beach, CA, USA: CVPR, 2019: 5873-5881
    [17]
    HE K, ZHANG X, REN S, et al. Deep residual learning for image recognition[C]//2016 IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Las Vegas, NV, USA, 2016: 770-778.
    [18]
    DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16x16 words: Transformers for image recognition at scale[EB/OL]. [2021-06-03][2025-03-10]. https://arxiv.org/abs/2010.11929.
    [19]
    TAN M, LE Q. Efficientnetv2: Smaller models and faster training[EB/OL].[2021-06-23][2025-03-10]. https://arxiv.org/abs/2104.00298.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索

    Figures(3)  / Tables(3)

    Article Metrics

    Article Views(14) PDF Downloads(1) Cited by()
    Proportional views
    Related
    Service
    Subscribe

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return