高级检索

    基于树皮纹理的轻量化YOLOv11树种识别方法

    Lightweight YOLOv11 tree species recognition method based on bark texture

    • 摘要:
      目的 为解决现有树种识别方法在多变光照条件下,因模型计算复杂度高而难以在硬件条件极端受限的移动端或边缘设备部署的问题,提出一种基于树皮纹理的轻量化树种识别方法。
      方法 本研究通过改进YOLOv11构建YOLOv11-SWER模型。首先,引入轻量化特征提取网络StarNet作为主干网络,结合深度可分离卷积与通道混洗机制,显著降低网络的参数量和特征提取过程中的计算量。其次,采用多分支特征融合模块RepNCSPELAN4,结合分组卷积与参数共享策略,兼顾全局特征与局部特征,提升多尺度特征融合效率。然后,设计小波池化(WaveletPool)层,减少噪声干扰并保留高频纹理细节,增强模型对树皮纹理微小特征的捕捉能力。最后,优化检测头结构Detect_Efficient,使用双分支分组卷积结构提高计算效率。同时,基于自建的70类树种、6681张树皮图像数据集,通过消融实验和对比实验对改进的模型性能进行充分的评估验证。
      结果 该模型的检测精确率、召回率、平均精度(mAP50)、平均精度均值(mAP50-95)以及精确率和召回率的调和平均数F1分数分别达到98.1%、98.4%、0.993、0.750和0.982,同时,相较于YOLOv11模型,其参数量和计算量分别降低46.92%和51.5%,大幅降低了模型的空间复杂度和计算复杂度。在不同光照场景下保持稳定的识别性能,展现出良好的光照鲁棒性。
      结论 本研究提出的YOLOv11-SWER模型通过轻量化设计与多尺度特征优化,在参数量减少近半的情况下,仍能保持高检测精度,实现了高检测精度与高效率的良好平衡。此方法有望应用于智能林业检测、城市林业资源管理等场景中。

       

      Abstract:
      Objective To address the issue of existing tree species identification methods being difficult to deploy on mobile or edge devices with extremely limited hardware due to high computational complexity under varying lighting conditions, this study proposes a lightweight tree species identification method based on bark texture.
      Method This research improves YOLOv11 to construct the YOLOv11-SWER model. First, the lightweight feature extraction network StarNet was introduced as the backbone network, combining depthwise separable convolution and channel shuffle mechanisms to significantly reduce the model’s parameter count and computational load during feature extraction. Second, a multi-branch feature fusion module, RepNCSPELAN4, was adopted, integrating group convolution and parameter-sharing strategies to balance global and local features, thereby enhancing multi-scale feature fusion efficiency. Then, a wavelet pooling (WaveletPool) layer was designed to reduce noise interference while preserving high-frequency texture details, improving the model’s ability to capture subtle bark texture features. Finally, the detection head structure Detect_Efficient was optimized using a dual-branch group convolution architecture to improve computational efficiency. Additionally, based on a self-built dataset of 70 tree species with 6 681 bark images, ablation studies and comparative experiments were conducted to thoroughly evaluate the performance of improved model.
      Result The model achieved a detection precision of 98.1%, recall of 98.4%, mean average precision (mAP50) of 0.993, mean average precision across different IoU thresholds (mAP50-95) of 0.750, and an F1 score (harmonic mean of precision and recall) of 0.982. Compared with the YOLOv11 model, the parameter count and computational load were reduced by 46.92% and 51.5%, respectively, significantly lowering the model’s spatial and computational complexity. It maintained stable identification performance under varying lighting conditions, demonstrating strong illumination robustness.
      Conclusion The proposed YOLOv11-SWER model, through lightweight design and multi-scale feature optimization, achieves high detection accuracy while reducing parameters by nearly half, striking a good balance between high accuracy and efficiency. This method holds promise for applications in intelligent forestry monitoring and urban forestry resource management.

       

    /

    返回文章
    返回