International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(1): 1-9
Published online March 25, 2024
https://doi.org/10.5391/IJFIS.2024.24.1.1
© The Korean Institute of Intelligent Systems
Xinzhi Hu1, Wang-Su Jeon2, Grezgorz Cielniak3, and Sang-Yong Rhee2
1Department of IT Convergence Engineering, Kyungnam University, Changwon, Korea
2Department of Computer Engineering, Kyungnam University, Changwon, Korea
3School Computer Science, University of Lincoln, Lincoln, UK
Correspondence to :
Sang-Yong Rhee (syrhee@kyungnam.ac.kr)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Sugar beet is a biennial herb with cold, drought, and salinity resistance and is one of the world’s major sugar crops. In addition to sugar, sugar beets are important raw materials for chemical and pharmaceutical products, and the residue after sugar extraction can be used to produce agricultural by-products, such as compound feed, which has a high comprehensive utilization value [1]. Field weeds, such as sugar beets, are harmful to crop growth and can compete with crops for sunlight and nutrients. If weeds are not removed in time during crop growth, they cause a decrease in crop yield and quality. Therefore, there is considerable interest in the development of automated machinery for selective weeding operations. The core component of this technology is a vision system that distinguishes between crops and weeds. To address the problems of difficult weed extraction, poor detection, and segmentation of region boundaries in traditional sugar beet detection, an end-to-end encoder–decoder model based on an improved UNet++ for segmentation is proposed in this paper and applied to sugar beet and weed detection. UNet++ can better fuse feature maps from different layers by skipping connections and can effectively preserve the details of sugar beet and weed images. The new model adds an attention mechanism to UNet++ by embedding the attention module into the upsampling process of UNet++ to suppress interference from extraneous noise. The improved model was evaluated on a sugar beet and weed dataset containing 1026 images. The image dataset in this study was obtained from sugar beet and weed images collected at the University of Bonn, Germany. According to the experimental results, the model can significantly eliminate noise and improve segmentation accuracy.
Keywords: ATT-NestedUNet, Deep learning, Weed detection, Semantic segmentation, Sugar beet
No potential conflict of interest relevant to this article was reported.
E-mail: huxinzhi0326@gmail.com
E-mail: jws2218@naver.com
E-mail: Grzegorz.Cielniak@gmail.com
E-mail: syrhee@kyungnam.ac.kr
International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(1): 1-9
Published online March 25, 2024 https://doi.org/10.5391/IJFIS.2024.24.1.1
Copyright © The Korean Institute of Intelligent Systems.
Xinzhi Hu1, Wang-Su Jeon2, Grezgorz Cielniak3, and Sang-Yong Rhee2
1Department of IT Convergence Engineering, Kyungnam University, Changwon, Korea
2Department of Computer Engineering, Kyungnam University, Changwon, Korea
3School Computer Science, University of Lincoln, Lincoln, UK
Correspondence to:Sang-Yong Rhee (syrhee@kyungnam.ac.kr)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Sugar beet is a biennial herb with cold, drought, and salinity resistance and is one of the world’s major sugar crops. In addition to sugar, sugar beets are important raw materials for chemical and pharmaceutical products, and the residue after sugar extraction can be used to produce agricultural by-products, such as compound feed, which has a high comprehensive utilization value [1]. Field weeds, such as sugar beets, are harmful to crop growth and can compete with crops for sunlight and nutrients. If weeds are not removed in time during crop growth, they cause a decrease in crop yield and quality. Therefore, there is considerable interest in the development of automated machinery for selective weeding operations. The core component of this technology is a vision system that distinguishes between crops and weeds. To address the problems of difficult weed extraction, poor detection, and segmentation of region boundaries in traditional sugar beet detection, an end-to-end encoder–decoder model based on an improved UNet++ for segmentation is proposed in this paper and applied to sugar beet and weed detection. UNet++ can better fuse feature maps from different layers by skipping connections and can effectively preserve the details of sugar beet and weed images. The new model adds an attention mechanism to UNet++ by embedding the attention module into the upsampling process of UNet++ to suppress interference from extraneous noise. The improved model was evaluated on a sugar beet and weed dataset containing 1026 images. The image dataset in this study was obtained from sugar beet and weed images collected at the University of Bonn, Germany. According to the experimental results, the model can significantly eliminate noise and improve segmentation accuracy.
Keywords: ATT-NestedUNet, Deep learning, Weed detection, Semantic segmentation, Sugar beet
The structure diagram of UNet.
The structure diagram of UNet++.
Structure diagram of CBAM.
Structure diagram of ATT-NestedUNet.
Four different depths ATT-NestedUNet structures.
Farmland information collection robot BoniRob.
Sample collection and labelingmap, e.g. sugar beet (green) weed (red).
Example instances from the dataset (a) weeds only, (b) beets only, (c) beets and weeds, (d) soil free of beets and weeds.
Experiment results with beets only.
Experiment results with weeds only.
Background-only experimental results.
Experiment results with beets, weeds, background.
Table 1 . MIOU values of L1, L2, L3, L4.
Model | MIOU |
---|---|
ATT-NestedUNet L1 | 90.03% |
ATT-NestedUNet L2 | 91.14% |
ATT-NestedUNet L3 | 91.43% |
ATT-NestedUNet L4 | 91.42% |
Table 2 . mIOU values for the three trained models.
Model | mIOU | Back ground | Weed | Sugar beet |
---|---|---|---|---|
UNet | 90.82% | 95.29% | 95.72% | 98.95% |
Nested UNet | 91.04% | 95.29% | 95.73% | 98.96% |
ATT-NestedUNet | 91.42% | 95.29% | 95.75% | 98.98% |
Gayoung Kim
International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(3): 287-294 https://doi.org/10.5391/IJFIS.2024.24.3.287Jeongmin Kim and Hyukdoo Choi
International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(2): 105-113 https://doi.org/10.5391/IJFIS.2024.24.2.105Herlawati Herlawati, Edi Abdurachman, Yaya Heryadi, and Haryono Soeparno
International Journal of Fuzzy Logic and Intelligent Systems 2023; 23(4): 389-398 https://doi.org/10.5391/IJFIS.2023.23.4.389The structure diagram of UNet.
|@|~(^,^)~|@|The structure diagram of UNet++.
|@|~(^,^)~|@|Structure diagram of CBAM.
|@|~(^,^)~|@|Structure diagram of ATT-NestedUNet.
|@|~(^,^)~|@|Four different depths ATT-NestedUNet structures.
|@|~(^,^)~|@|Farmland information collection robot BoniRob.
|@|~(^,^)~|@|Sample collection and labelingmap, e.g. sugar beet (green) weed (red).
|@|~(^,^)~|@|Example instances from the dataset (a) weeds only, (b) beets only, (c) beets and weeds, (d) soil free of beets and weeds.
|@|~(^,^)~|@|Experiment results with beets only.
|@|~(^,^)~|@|Experiment results with weeds only.
|@|~(^,^)~|@|Background-only experimental results.
|@|~(^,^)~|@|Experiment results with beets, weeds, background.