Real time weed identification with enhanced mobilevit model for mobile devices
Abstract Deep learning model optimization have notably enhanced weed identification accuracy. However, there is a short-fall in detailed research on optimizing models for weed identification with images from mobile embedded systems. Also, existing methods generally use large, slow multi-layer convol...
Saved in:
| Main Authors: | , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Nature Portfolio
2025-07-01
|
| Series: | Scientific Reports |
| Subjects: | |
| Online Access: | https://doi.org/10.1038/s41598-025-12036-0 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Abstract Deep learning model optimization have notably enhanced weed identification accuracy. However, there is a short-fall in detailed research on optimizing models for weed identification with images from mobile embedded systems. Also, existing methods generally use large, slow multi-layer convolutional networks (CNNs), which are impractical for use on mobile embedded devices. To address those issues, we propose a lightweight weed identification model based on an enhanced MobileViT architecture, effectively balancing high accuracy with real-time performance. Our approach begins with the application of a multi-scale retinal enhancement algorithm featuring color restoration to preprocess image data. This step improves the clarity of images, particularly those with blurred edges or significant shadow interference. Following this, we introduce an optimized MobileViT model that incorporates the Efficient Channel Attention (ECA) module into the weed feature extraction network. This design ensures robust feature extraction capabilities while simultaneously reducing the model’s parameters and computational complexity. The MobileViT model within our feature extraction network is engineered to concurrently learn local and global semantic information. This capability allows it to accurately distinguish subtle differences between weeds and crops by leveraging a minimal number of modules. To demonstrate the effectiveness of our model, it achieved an F1 score of 98.51% and an average identification time of 89 milliseconds per image. These results underscore its suitability for lightweight deployment, maintaining high accuracy while minimizing model complexity. |
|---|---|
| ISSN: | 2045-2322 |