Article Search
닫기

Original Article

Split Viewer

International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(1): 1-9

Published online March 25, 2024

https://doi.org/10.5391/IJFIS.2024.24.1.1

© The Korean Institute of Intelligent Systems

ATT-NestedUnet: Sugar Beet and Weed Detection Using Semantic Segmentation

Xinzhi Hu1, Wang-Su Jeon2, Grezgorz Cielniak3, and Sang-Yong Rhee2

1Department of IT Convergence Engineering, Kyungnam University, Changwon, Korea
2Department of Computer Engineering, Kyungnam University, Changwon, Korea
3School Computer Science, University of Lincoln, Lincoln, UK

Correspondence to :
Sang-Yong Rhee (syrhee@kyungnam.ac.kr)

Received: April 17, 2023; Accepted: March 20, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Sugar beet is a biennial herb with cold, drought, and salinity resistance and is one of the world’s major sugar crops. In addition to sugar, sugar beets are important raw materials for chemical and pharmaceutical products, and the residue after sugar extraction can be used to produce agricultural by-products, such as compound feed, which has a high comprehensive utilization value [1]. Field weeds, such as sugar beets, are harmful to crop growth and can compete with crops for sunlight and nutrients. If weeds are not removed in time during crop growth, they cause a decrease in crop yield and quality. Therefore, there is considerable interest in the development of automated machinery for selective weeding operations. The core component of this technology is a vision system that distinguishes between crops and weeds. To address the problems of difficult weed extraction, poor detection, and segmentation of region boundaries in traditional sugar beet detection, an end-to-end encoder–decoder model based on an improved UNet++ for segmentation is proposed in this paper and applied to sugar beet and weed detection. UNet++ can better fuse feature maps from different layers by skipping connections and can effectively preserve the details of sugar beet and weed images. The new model adds an attention mechanism to UNet++ by embedding the attention module into the upsampling process of UNet++ to suppress interference from extraneous noise. The improved model was evaluated on a sugar beet and weed dataset containing 1026 images. The image dataset in this study was obtained from sugar beet and weed images collected at the University of Bonn, Germany. According to the experimental results, the model can significantly eliminate noise and improve segmentation accuracy.

Keywords: ATT-NestedUNet, Deep learning, Weed detection, Semantic segmentation, Sugar beet

This result was supported by “Regional Innovation Strategy (RIS)” through the National Research Foundation of Korea (NRF), funded by the Ministry of Education (MOE) (No. 2021RIS-003).

No potential conflict of interest relevant to this article was reported.

Xin-Zhi Hu received her B.S. in Computer Engineering from Kyungnam University, Masan, South Korea in 2021 and is currently pursuing the M.S. in IT Convergence Engineering at Kyungnam University, Masan, South Korea. Her present interests include computer vision pattern recognition.

E-mail: huxinzhi0326@gmail.com

Wang-Su Jeon received his B.S. and M.S. degrees in Computer Engineering and IT Convergence Engineering from Kyungnam University, Masan, South Korea, in 2016 and 2018, and is currently pursuing the Ph.D. in IT Convergence Engineering at Kyungnam University, Masan, South Korea. His present interests include computer vision, pattern recognition, and machine learning.

E-mail: jws2218@naver.com

Grzegorz Cielniak is a senior lecturer at the School of Computer Science, University of Lincoln. He received his M.Sc. in Robotics from the Wrocław University of Technology in 2000 and Ph.D. in Computer Science from the Örebro University in 2006. His research interests include mobile robotics, artificial intelligence, real-time computer vision systems and multisensor fusion with particular focus on robotic applications in agriculture.

E-mail: Grzegorz.Cielniak@gmail.com

Sang-Yong Rhee received his B.S. and M.S. degrees in Industrial Engineering from Korea University, Seoul, South Korea, in 1982 and 1984, respectively, and his Ph.D. degree in Industrial Engineering at Pohang University, Pohang, South Korea. He is currently a professor at the Computer Engineering, Kyungnam University, Masan, South Korea. His research interests include computer vision, augmented reality, neuro-fuzzy and human-robot interface.

E-mail: syrhee@kyungnam.ac.kr

Article

Original Article

International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(1): 1-9

Published online March 25, 2024 https://doi.org/10.5391/IJFIS.2024.24.1.1

Copyright © The Korean Institute of Intelligent Systems.

ATT-NestedUnet: Sugar Beet and Weed Detection Using Semantic Segmentation

Xinzhi Hu1, Wang-Su Jeon2, Grezgorz Cielniak3, and Sang-Yong Rhee2

1Department of IT Convergence Engineering, Kyungnam University, Changwon, Korea
2Department of Computer Engineering, Kyungnam University, Changwon, Korea
3School Computer Science, University of Lincoln, Lincoln, UK

Correspondence to:Sang-Yong Rhee (syrhee@kyungnam.ac.kr)

Received: April 17, 2023; Accepted: March 20, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Sugar beet is a biennial herb with cold, drought, and salinity resistance and is one of the world’s major sugar crops. In addition to sugar, sugar beets are important raw materials for chemical and pharmaceutical products, and the residue after sugar extraction can be used to produce agricultural by-products, such as compound feed, which has a high comprehensive utilization value [1]. Field weeds, such as sugar beets, are harmful to crop growth and can compete with crops for sunlight and nutrients. If weeds are not removed in time during crop growth, they cause a decrease in crop yield and quality. Therefore, there is considerable interest in the development of automated machinery for selective weeding operations. The core component of this technology is a vision system that distinguishes between crops and weeds. To address the problems of difficult weed extraction, poor detection, and segmentation of region boundaries in traditional sugar beet detection, an end-to-end encoder–decoder model based on an improved UNet++ for segmentation is proposed in this paper and applied to sugar beet and weed detection. UNet++ can better fuse feature maps from different layers by skipping connections and can effectively preserve the details of sugar beet and weed images. The new model adds an attention mechanism to UNet++ by embedding the attention module into the upsampling process of UNet++ to suppress interference from extraneous noise. The improved model was evaluated on a sugar beet and weed dataset containing 1026 images. The image dataset in this study was obtained from sugar beet and weed images collected at the University of Bonn, Germany. According to the experimental results, the model can significantly eliminate noise and improve segmentation accuracy.

Keywords: ATT-NestedUNet, Deep learning, Weed detection, Semantic segmentation, Sugar beet

Fig 1.

Figure 1.

The structure diagram of UNet.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 2.

Figure 2.

The structure diagram of UNet++.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 3.

Figure 3.

Structure diagram of CBAM.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 4.

Figure 4.

Structure diagram of ATT-NestedUNet.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 5.

Figure 5.

Four different depths ATT-NestedUNet structures.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 6.

Figure 6.

Farmland information collection robot BoniRob.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 7.

Figure 7.

Sample collection and labelingmap, e.g. sugar beet (green) weed (red).

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 8.

Figure 8.

Example instances from the dataset (a) weeds only, (b) beets only, (c) beets and weeds, (d) soil free of beets and weeds.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 9.

Figure 9.

Experiment results with beets only.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 10.

Figure 10.

Experiment results with weeds only.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 11.

Figure 11.

Background-only experimental results.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Fig 12.

Figure 12.

Experiment results with beets, weeds, background.

The International Journal of Fuzzy Logic and Intelligent Systems 2024; 24: 1-9https://doi.org/10.5391/IJFIS.2024.24.1.1

Table 1 . MIOU values of L1, L2, L3, L4.

ModelMIOU
ATT-NestedUNet L190.03%
ATT-NestedUNet L291.14%
ATT-NestedUNet L391.43%
ATT-NestedUNet L491.42%

Table 2 . mIOU values for the three trained models.

ModelmIOUBack groundWeedSugar beet
UNet90.82%95.29%95.72%98.95%
Nested UNet91.04%95.29%95.73%98.96%
ATT-NestedUNet91.42%95.29%95.75%98.98%

Share this article on :

Related articles in IJFIS