A CNN Based Model To Count The Leaves of Rosette Plants (LC Net)
A CNN Based Model To Count The Leaves of Rosette Plants (LC Net)
com/scientificreports
Plant image analysis is a significant tool for plant phenotyping. Image analysis has been used to
assess plant trails, forecast plant growth, and offer geographical information about images. The area
segmentation and counting of the leaf is a major component of plant phenotyping, which can be used
to measure the growth of the plant. Therefore, this paper developed a convolutional neural network-
based leaf counting model called LC-Net. The original plant image and segmented leaf parts are fed as
input because the segmented leaf part provides additional information to the proposed LC-Net. The
well-known SegNet model has been utilised to obtain segmented leaf parts because it outperforms
four other popular Convolutional Neural Network (CNN) models, namely DeepLab V3+, Fast FCN
with Pyramid Scene Parsing (PSP), U-Net, and Refine Net. The proposed LC-Net is compared to the
other recent CNN-based leaf counting models over the combined Computer Vision Problems in Plant
Phenotyping (CVPPP) and KOMATSUNA datasets. The subjective and numerical evaluations of the
experimental results demonstrate the superiority of the LC-Net to other tested models.
Recently, plant phenomics has attracted the rising attention of researchers. There is a need to increase the
capabilities of high-yielding plants to ensure our food security. Global agriculture has faced significant chal-
lenges, including the need for high-yielding plants that can adapt to future climates and identifying the specific
feedstock crop for biofuel p roduction1. The nutrients, the supply of carbon, and the other outer environmental
lant2. Plant phenotyping can provide an understanding
factors affect the size of leaves as well as the growth of a p
of plant genes, correlations between plants and the environment, plant trails, etc. as well as new technologies
to increase the yield of plants to overcome those problems. Plant phenotyping is also useful for studying the
growth of a plant, the yielding rate of a plant, and also the internal structure of a plant. The traditional laborious
and costly plant phenotyping techniques have become a bottleneck in plant breeding techniques as well as in
functional genomics1. These phenotyping bottlenecks have prevented us from understanding the correlation
between expressed phenotypes, genetic factors, and the condition of the e nvironment3.
The possibility of connecting heritable traits to genetic markers depends on precise phenotypic a ssessment4.
Plant image analysis is also a valuable tool for plant phenotyping. Image analysis has been used to analyse the
plant trails and predict the growth of the plant, as well as give spatial information about the image. But meas-
uring the visual characteristics of a plant is very costly and needs a very detailed long-term investigation to
continue the observation. As a result, an automated approach for resolving the problem must be established. As
recent literature suggests, the deep-learning (DL)-based methods are one of the most contemporary artificial-
intelligence (AI) techniques and are currently a crucial component of plant phenotyping. There is a survey on
plant phenotyping using computer vision with D L 5.
Leaf area segmentation is the key feature of plant phenotyping and can be used to analyze the growth of a
plant2. However, the segmentation of the leaves has become challenging when the size of the leaves is tiny or a
1
Wipro Technologies, Pune, Maharashtra, India. 2Department of Computer Science and Application, Midnapore
College (Autonomous), Paschim Medinipur, West Bengal, India. 3Department of Computer and Information
Science, Linköping University, Linköping, Sweden. 4Faculty of Science, Fayoum University, Fayoum, Egypt. 5MEU
Research Unit, Middle East University, Amman, Jordan. 6Applied Science Research Center, Applied Science
Private University, Amman 11931, Jordan. 7Hourani Center for Applied Scientific Research, Al-Ahliyya Amman
University, Amman 19328, Jordan. 8Computer Science Department, Al al-Bayt University, 25113, Mafraq,
Jordan. 9Artificial Intelligence and Sensing Technologies (AIST) Research Center, University of Tabuk, 71491, Tabuk,
Saudi Arabia. 10Department of Electrical and Computer Engineering, Lebanese American University, 13‑5053,
Byblos, Lebanon. 11College of Engineering, Yuan Ze University, Taoyuan, Taiwan. 12Department of Computer
Science and Engineering, Indian Institute of Technology, Delhi, India. *email: [email protected]
Vol.:(0123456789)
www.nature.com/scientificreports/
significant number of leaves are overlapped. Also, the angle of the images and the lighting effect on the leaves
can affect the efficiency of leaf area segmentation. Another crucial factor for analysing plant growth in plant
phenotyping is determining the leaf count of the plant6. It is also challenging and time-consuming process to
determine the number of leaves. A comprehensive plant phenotyping method for camera-captured images
contributes to the cost reduction and improvement of plant and agricultural production. As a result, numerous
researchers are looking into plant phenotyping using images recorded by cameras.
Furthermore, current CNN-based methods show their efficient performance in accurate leaf count predic-
tions. A CNN is a class of neural networks within the field of DL. CNNs are designed with one or more convo-
lutional layers and are mostly used for tasks such as image processing, classification, segmentation, and analysis
of auto-correlated data. CNNs are primarily employed for the purpose of extracting features from grid-like
matrix datasets, particularly in the context of image analysis. The CNN architecture encompasses various levels,
including the input layer, convolutional layer, activation layer, pooling layer, and completely linked layers. The
convolutional layer is responsible for applying filters to the input image in order to extract relevant features.
The activation function is then applied in an element-wise manner to modify the matrix values, specifically
converting negative values to zero. Following this, the pooling layer is employed to downsample the image,
thereby reducing computational requirements. Finally, the fully connected layer is used to make the ultimate
prediction. The process by which the network acquires the most effective filters is achieved through the use of
backpropagation and gradient descent.
Therefore, in this manuscript, a novel CNN-based model for leaf counting has been proposed called LC-Net.
The proposed model takes segmented leaf parts as additional input for achieving better leaf counting accuracy.
Hence, well-established SegNet7 is utilized as it performs better leaf segmentation compared to existing well-
established CNN-based models, namely, DeepLab V3+8, U-Net9, Fast FCN with Pyramid Scene Parsing (PSP)10,
and Refine N et11. Lastly, the proposed LC-Net has been compared with the other state-of-the-art leaf count
models, and the performance of all tested leaf count models has been tested over the combined CVPPP and
KOMATSUNA datasets. The qualitative and numerical results indicate that the proposed LC-Net outperforms
the existing leaf counting results.
In a nutshell, the significant contributions of the proposed work are enlisted in the following:
– A novel CNN-based model to count the number of rosette plants’ leaves has been proposed which has been
provided two inputs i.e., segmented output and the original image for better accuracy
– The proposed LC-Net model is tested on the merged version of the dataset of KOMATSUNA and CVPPP
(annotated for the leaf counting task by the experts). It is seen that the proposed LC-Net model outperforms
existing state-of-the-art techniques.
– The proposed model use a normalization layer that is used to filter the unwanted pixels present in the images.
Application of this layer in the proposed LC-Net is discussed in detail in Sect. "CNN based leaf segmentation".
The remainder of this manuscript is structured as follows. Next, Sect. "Related work" summarises about the exist-
ing work in the related field. Thereafter, Sect. "Methodology" discusses the proposed technique. Furthermore,
Sect. "Experimental results" presents the experimental investigation. Lastly, Sect. "Conclusion and future work"
contains the final observations and concluding remarks.
Related work
The past few years have witnessed successful research work in the domain of phenotyping of plants. In the recent
past, several research papers have been published on this hot topic. For example, in Ref.12, Tu et al. developed
a leaf counting network architecture based on YOLO V3. They considered the problem of leaf counting as an
object detection problem. As a result, first they detected the leaves and drew the bounding boxes, and then they
counted the number of boxes. The proposed model achieved -0.32 of difference in count (DiC), 0.48 of Absolute
difference in count (AbsDiC), 0.80 of Mean Squared Error (MSE), and 64.00% of percentage agreement on the
CVPPP dataset. In Ref.13, another deep-learning based model called Eff-U-Net++ for leaf segmentation and
counting had been proposed by Bhagat et al. Eff-U-Net++ was an encoder-decoder-based network. It had applied
EfficientNet-B4 as the encoder sub-net. The model achieved 0.11, 0.03, 0.12 DiC and 0.21, 0.38, 1.27 AbsDiC on
the CVPPP14, MSU-PID15, and KOMATSUNA16 datasets, respectively.
In Ref.17, two novel DL approaches had been proposed, elaborated, and compared by Farjon et al. for visual
leaf counting tasks. The first model fed input image in different resolutions into the network that had ResNet-50
as the backbone network to estimate features of leaves at multiple scales. Next, in order to get the final count,
repetitive regression of leaf heat map got from the previous step. The second model counted the number of leaves
by locating the centers of the detected leaves and finally aggregating them. They got 1.17 MSE and 43.7% of the
percentage agreement on the CVPPP dataset14. Miao et al.18 proposed a dataset for maize leaf counting and also
proposed two DL methods to count the maize leaves. The first model counted the leaves by regression, and the
second model counted the leaves by detection. In Ref.19, Lu et al. came up with a better way to use DL to count
the dense leaves and leaves that overlap in natural environments. The developed approach was used to detect
the object by merging a space-to-depth module, an Atrous spatial pyramid pooling, and a convolutional block
attention module into the network. The experimental results showed that the improved DL approach achieved
96% accuracy. To improve the phenotyping process, in Ref.20, Karthik et al. introduced a unique semantic seg-
mentation pipeline for the segmentation task. In CVPPP14 competition, a number of deep CNN models such
as U-Net, Attention-Augmented Net, and Attention-Net were introduced. These networks were trained using
the Arabidopsis Thaliana plant dataset. The Attention-Net achieved a 0.985 dice score, which is the best among
others.
Vol:.(1234567890)
www.nature.com/scientificreports/
In Ref.21, Kumar et al. came up with a new orthogonal transform domain-based method to segment the leaf
region and further counted it by fine-tuned deep CNN models. On the CVPPP dataset, fine-tuned AlexNet
and VGG19 had been used to count the leaves and got 25.51%, 33.67% of percentage agreement, 5.43, 2.03 of
MSE, 1.71, 1.03 of AbsDiC, and 0.39, 0.11 of DiC, respectively. In Ref.22, Buzzy et al. proposed a novel real-time
object detection approach for the identification, localization, and quantification of plant leaves. A comparative
analysis was conducted between a Tiny-YOLOv3 model and a faster R-CNN model. The Tiny-YOLOv3 and
Faster R-CNN models were evaluated based on several performance metrics, including DiC, AbsDiC, MSE, and
percentage agreement. The obtained values for these metrics were 0.25 and 0.0556, 0.8056 and 1.2778, 2.0833
and 2.8889, and 56% and 27.78%, respectively. Hati et al.23 presented a regression model for the purpose of leaf
counting. The images were subjected to segmentation and subsequent enhancement, resulting in the removal
of noise and transformation of the pixel data associated with the leaves. Subsequently, the images were inputted
into the regression model, which is founded upon the architecture of AlexNet. In Ref.24, Ayalew et al. presented
a domain-adversarial learning method in which a domain adaption technique was used to estimate a density
map for leaf counting. Due to its flexibility in accommodating variations in distribution across source and des-
tination datasets, the method exhibits potential for application in a broader spectrum of leaf counting and plant
organ counting scenarios. The method got -0.95 of DiC, 1.56 of AbsDiC, 5.26 of MSE, and 29.33% of percentage
agreement on the CVPPP dataset. Gomes and Zheng25 presented an experimental study on the limitations of
datasets used for phenotyping and the performance strategy of the leaf segmentation tasks. They also looked at
how test-time augmentation and model cardinality might help with single-class image segmentation. Another
investigation had been conducted in Ref.26 by Yang et al. where they utilized a mask R-CNN based model to
effectively segregate and classify leaf images that contained intricate backgrounds.
In the year 2019, a new a pproach27 to extract leaf regions from images of plants and count the number of leaves
had been introduced by Kumar et al. There were three phases to the proposed methodology which are statistical
image enhancing technique, a graph-based leaf area extraction approach, and a circular hough transform (CHT)
based technique. In Ref.28, Valente et al. presented a preliminary study showcasing the efficacy of a trained deep
neural network in accurately quantifying the number of leaves in plant images obtained by greenhouse workers
through the use of handheld equipment. They got 0.31 of DiC, 0.62 of AbsDiC, 0.77 of MSE, and 47% of percent-
age Agreement. A Google Inception Net V3 based CNN model had been used by Jiang et al. in Ref.29 to count the
number of maize leaves. To reduce redundant information, the Fisher Vector (FV) was used, and the Random
Forest (RF) method was used to get the final prediction. They got 0.0018 of DiC, 0.35 of AbsDiC, and 0.31 of MSE.
In 2018, Giuffrida et al. presented a single deep network30 that counted the number of leaves from multi-
modal 2D images of different species for any rosette-shaped plant. The model had a DiC of 0.19, an AbsDiC of
0.91, a MSE of 1.56, and a 32.9 percentage agreement. For the CVPPP dataset, the proposed approach achieved
a segmentation accuracy of 95.4%, a DiC of 0.7 and an AbsDiC of 2.3. In Ref.31, Ubbens et al. proposed a plant
phenotyping dataset by demonstrating that performance could be improved on the leaf counting task using 3D
synthetic plants. These synthetic plants were also applied to augment a dataset. They recreated the architecture
used in the reference experiment using the Ara2013-Canon dataset in the augmentation experiment3 for the
augmentation experiment. They also switched the real and synthetic image datasets utilized to both train and
test the DL-based model. Additionally, it was demonstrated that these datasets could be utilised in a comparable
manner for training a neural network to accurately quantify the number of leaves.
For the plant phenotyping task, Aich et al. developed a data-driven strategy in Ref.32 that could be employed
to a variety of plant species and imaging configurations. They used a deconvolutional network to segment the
leaves, and the predicted images were used in a CNN for leaf counting. The model got 0.73 of DiC, 1.62 of Abs-
DiC, 4.31 of MSE, and 24.0 of Percentage Agreement on the CVPPP dataset. A Data Augmentation technique
had been proposed by Kuznichov et al. in Ref.33 for segmenting the leaves followed by the counting the leaves
from the images of rosette plants. In Ref.34, Itzhaky et al. developed two novel deep learning algorithms designed
for the purpose of leaf item counting. The researchers utilised the CVPPP 2017 Leaf Counting Challenge dataset
to demonstrate the efficacy of these methods. The findings show that they defeated the CVPPP challenge winner
from 2017. In Ref.35, Pape and Kulkas introduced a methodology for leaf segmentation that utilises edge detection
techniques. Additionally, they devised a method to analyse images by employing the software IAP36 in order to
extract a multitude of image attributes that might be utilised for estimating the quantity of leaves.
The aforementioned discourse unequivocally illustrates the immense utility of performing precise tasks,
such as leaf segmentation and counting, in the field of plant phenotyping. Furthermore, it should be noted that
there exists a restricted range of CNN models and techniques that have demonstrated the capability to generate
precise outcomes across two widely utilised datasets, namely CVPPP and KOMATSUNA. It has been observed
that these algorithms frequently encounter difficulties in scenarios including a green background and/or a sub-
stantial amount of overlapping leaves. Consequently, in order to address this deficiency, the primary objective
of this research endeavour is to construct a Convolutional Neural Network (CNN) model, specifically referred
to as LC-Net. The LC-Net, as proposed, demonstrates the ability to achieve higher predictions in leaf counting
even when confronted with various types of backgrounds. Furthermore, LC-Net demonstrates the capability
to precisely quantify the number of leaves, even in cases when they are overlapping. The next part provides a
detailed description of the LC-Net model that has been proposed, as well as an overview of the dataset that was
utilised in this study.
Vol.:(0123456789)
www.nature.com/scientificreports/
Methodology
The proposed LC-Net has been designed to count the number of leaves in the rosette plants. The suggested model
takes both the RGB image of a rosette plant and the segmented image of its leaves as input in order to enhance
accuracy. The subsequent subsections provide an exposition of the LC-Net architecture and an elucidation of
the design’s underlying reasoning.
Normalization layer
The normalization layer has been utilized to improve the segmented outcomes. This normalization Layer’s pur-
pose is to eliminate any unwanted pixels from the images. The uneven background or light reflection could be
the cause of these unwanted pixels. The precision of segmentation and leaf count may be affected as a result of
these circumstances. The inclusion of the normalization layer has facilitated the construction of models that
exhibit enhanced predictive accuracy. Figure 1 shows the segmented outcomes of SegNet with and without using
normalization layer. Visual analysis clearly shows that utilization of normalization layer over SegNet’s output
enhances the results.
Experimental results
The model is trained and tested using a system containing an NVIDIA GeForce 1650 having cuDNN CUDA
10.0, a 256 GB SSD, 16 GB RAM, and an AMD Ryzen 5 3550H CPU. TensorFlow and Scikitlearn were used in
all of the studies. The experimental results have two folds: one is leaf segmentation and other is leaf counting.
Both the results are presented in the following sub-sections.
Dataset design
As stated earlier, experiments on both the leaf segmentation and leaf counting have been done on two well-known
datasets. They are (i) the KOMATSUNA [14] dataset, and (ii) the plant phenotyping benchmark dataset, known
as CVPPP14. In our experiment, these datasets are merged. The KOMATSUNA dataset has been annotated for
the leaf counting task by the experts. The CVPPP dataset have four sections. They are named as A1, A2, A3, and
A4. All sections are included with segmented ground truth. On the other hand, there are two portions in the the
Vol:.(1234567890)
www.nature.com/scientificreports/
KOMATSUNA dataset. Among them, one portion is made by capturing the images by a RGB-D camera, while
the images in the other segment of the dataset is captured by several RGB cameras. As a whole, there are total of
1200 images in these two categories. As a result, there are a total of 2010 images along with the corresponding
segmentation and leaf number ground truths in the the combined dataset. Bilinear i nterpolation41 has been used
to resize the images into 224×224. Subsequently, the combined dataset is partitioned into distinct subsets, namely
the training, validation, and testing sets of images. In our experiment, the training set consists of 1410 images,
whereas the validation and test sets each contain 300 images. In addition, a series of vertical flips, 90◦ and 180◦
clockwise and anticlockwise rotations are randomly applied in order to generate a greater quantity of images. In
Fig. 3, some of the samples from each of the dataset are shown. In order to compare more accurately with state-
of-the-art CNN models, another dataset is prepared using 160 images chosen randomly from the CVPPP dataset.
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 2. (a) Conv Block, (b) Architecture of the proposed LC-Net model.
has been used as the loss function. In our experiment, the size of the input and output images are size 224 x 224
x 3 and 224 x 224 x 1, respectively. The batch size varies depending on the machine capacity and the architecture
of the applied model. The suggested and existing tested CNN-based models’ parameter settings are provided
in Table 1.
In order to evaluate and compare the performance of the segmentation process by the CNN models, three
commonly used metrics such as, (a) segmentation accuracy (AC); (b) intersection over union (IoU); and (c) dice
score (DI), are used. The quality metrics are summarised in Table 2.
In the training period, DeepLab V3+ shows the best performance among all. DeepLab V3+ achieves 97.38%
of train accuracy, 97.30% of test accuracy, and 0.0108 of train loss. But while the models have been tested on the
merged dataset and on the CVPPP dataset, the SegNet model shows the best results. Test results on the merged
and CVPPP datasets are shown in Tables 3 and 4, respectively. SegNet achieves a 95.04% dice score and a 90.58%
IoU. Therefore, the SegNet model has been chosen as the segmentation model for the proposed leaf counting
Vol:.(1234567890)
www.nature.com/scientificreports/
Table 1. Parameters Setting of the existing CNN-based models used for leaf segmentation.
Models Test loss (×10−3 ) Test accuracy (%) IoU Dice score
DeepLab V3+ 22.3 97.06 89.78 94.59
SegNet 17.6 97.29 90.58 95.04
Fast FCN with PSP 19.4 97.24 89.29 94.31
U-Net 58.0 96.40 86.50 91.87
Refine Net 222.0 97.10 89.23 94.27
Table 3. Test result that are obtained by various architectures on merged dataset (Test set size = 300 images).
Significant values are in bold.
Models Test loss (×10−3 ) Test accuracy (%) IoU Dice score
DeepLab V3+ 23.7 97.21 85.11 91.79
SegNet 19.7 97.92 86.04 92.50
Fast FCN with PSP 21.1 97.57 85.94 91.54
U-Net 59.8 96.92 81.55 89.23
Refine Net 24.2 97.06 84.50 91.20
Table 4. Test result that are obtained by various CNN model on CVPPP Dataset (Test set size = 160 images).
Significant values are in bold.
model, i.e., LC-Net. The outcomes of segmentation by the existing tested CNN-based models along with the
ground truth on the merged dataset are depicted in Fig. 4. From Fig. 4, it is quite clear that the SegNet model is
performing with a significant accuracy.
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 4. Qualitative comparison of the segmentation of proposed method with existing approaches.
Four commonly used metrics namely28,21, (a) Mean Square Error (MSE); (b) Abstract DiC; (c) R2; and (d)
Percentage Agreement are utilized to study the performance of Leaf counting models. The brief descriptions of
the mentioned metrics are reported in Table 6.
After training, we have tested all models on our merged test dataset and the CVPPP dataset as well. We
have used four evolution measures, i.e., Abs DiC, MSE, R2, percentage Agreement (%) and testing loss. Table 7
shows the test results of the models on the merged dataset, and Table 8 shows the test results of the models on
Vol:.(1234567890)
www.nature.com/scientificreports/
Table 7. Testing result of leaf count have obtained by different architectures on merged dataset (Test set size =
300 images). Significant values are in bold.
Table 8. Testing result of leaf count have obtained by different architectures on CVPPP dataset (Test set size =
160 images). Significant values are in bold.
the CVPPP dataset. The test results clearly show that the proposed LC-Net model is better than all other tested
models and methods on both datasets by a large margin.
Vol.:(0123456789)
www.nature.com/scientificreports/
Table 9. Numerical result for measuring the effect of combined input over CVPPP dataset (Test set size = 160
images). Significant values are in bold.
LC-Net with original image as input over CVPPP datasets. It can be seen that when the segmented image is
combined with the original image the performance of the LC-Net model has been enhanced to some great extent.
Best outputs are highlighted in bold in the tables. The presence of the ↑ symbol following a measure signifies
that a higher value of that metric indicates a more favourable outcome. In contrast, the presence of the sign ↓
following the metric denotes the opposite.
Data availibility
The datasets that support the findings of this study are publicly available. Link for CVPPP dataset is: http://
www.plant-phenotyping.org/datasets. Link for KOMATSUNA dataset is: https://limu.ait.kyushu-u.ac.jp/ agri/
komatsuna/.
References
1. Furbank, R. T. & Tester, M. Phenomics-technologies to relieve the phenotyping bottleneck. Trends Plant Sci. 16(12), 635–644
(2011).
2. Walter, A. & Schurr, U. The modular character of growth in Nicotiana tabacum plants under steady-state nutrition. J. Exp. Bot.
50(336), 1169–1177 (1999).
3. Ubbens, J. R. & Stavness, I. Deep plant phenomics: A deep learning platform for complex plant phenotyping tasks. Front. Plant
Sci. 8, 1190 (2017).
4. Jin, S. et al. Stem-leaf segmentation and phenotypic trait extraction of individual maize using terrestrial LiDAR data. IEEE Trans.
Geosci. Remote Sens. 57(3), 1336–1346 (2018).
5. Chandra, A. L., Desai, S. V., Guo, W., & Balasubramanian, V. N. Computer vision with deep learning for plant phenotyping in
agriculture: A survey. Preprint at arXiv:2006.11391. (2020).
6. Telfer, A., Bollman, K. M. & Poethig, R. S. Phase change and the regulation of trichome distribution in Arabidopsis thaliana.
Development 124(3), 645–654 (1997).
7. Badrinarayanan, V., Kendall, A. & Cipolla, R. SegNet: A deep convolutional encoder-decoder architecture for image segmentation.
IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481–2495 (2017).
8. Chen, L. C., Zhu, Y., Papandreou, G., Schroff, F., & Adam, H. Encoder-decoder with atrous separable convolution for semantic
image segmentation. In Proceedings of the European conference on computer vision (ECCV) (pp. 801-818) (2018).
9. Ronneberger, O., Fischer, P. & Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Confer-
ence on Medical Image Computing and Computer-Assisted Intervention (eds Ronneberger, O. et al.) 234–241 (Springer, 2015).
10. Wu, H., Zhang, J., Huang, K., Liang, K., & Yu, Y. Fastfcn: Rethinking dilated convolution in the backbone for semantic segmenta-
tion. Preprint at arXiv:1903.11816. (2019).
11. Lin, G., Milan, A., Shen, C., & Reid, I. Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In
Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1925-1934) (2017).
12. Tu, Y. L., Lin, W. Y., & Lin, Y. C. Toward automatic plant phenotyping: starting from leaf counting. Multimedia Tools and Applica-
tions, 1-15 (2022).
13. Bhagat, S., Kokare, M., Haswani, V., Hambarde, P. & Kamble, R. Eff-UNet++: A novel architecture for plant leaf segmentation and
counting. Eco. Inform. 68, 101583 (2022).
Vol:.(1234567890)
www.nature.com/scientificreports/
14. Minervini, M., Fischbach, A., Scharr, H. & Tsaftaris, S. A. Finely-grained annotated datasets for image-based plant phenotyping.
Pattern Recogn. Lett. 81, 80–89 (2016).
15. Cruz, J. A. et al. Multi-modality imagery database for plant phenotyping. Mach. Vis. Appl. 27(5), 735–749 (2016).
16. Uchiyama, H., Sakurai, S., Mishima, M., Arita, D., Okayasu, T., Shimada, A., & Taniguchi, R. I. An easy-to-setup 3D phenotyping
platform for KOMATSUNA dataset. In Proceedings of the IEEE International Conference on Computer Vision Workshops (pp.
2038-2045) (2017).
17. Farjon, G., Itzhaky, Y., Khoroshevsky, F. & Bar-Hillel, A. Leaf counting: Fusing network components for improved accuracy. Front.
Plant Sci. 12, 1063 (2021).
18. Miao, C. et al. Automation of leaf counting in maize and sorghum using deep learning. Plant Phenome J. 4(1), e20022 (2021).
19. Lu, S. et al. Counting dense leaves under natural environments via an improved deep-learning-based object detection algorithm.
Agriculture 11(10), 1003 (2021).
20. Karthik, P., Parashar, M., Reka, S. S., Rajamani, K. T. & Heinrich, M. P. Semantic segmentation for plant phenotyping using
advanced deep learning pipelines. Multimedia Tools Appl. 81(3), 4535–4547 (2022).
21. Kumar, J. P. & Domnic, S. Rosette plant segmentation with leaf count using orthogonal transform and deep convolutional neural
network. Mach. Vis. Appl. 31(1), 1–14 (2020).
22. Buzzy, M., Thesma, V., Davoodi, M. & MohammadpourVelni, J. Real-time plant leaf counting using deep object detection networks.
Sensors 20(23), 6896 (2020).
23. Hati, A. J. & Singh, R. R. Smart indoor farms: Leveraging technological advancements to power a sustainable agricultural revolu-
tion. AgriEngineering 3(4), 728–767 (2021).
24. Ayalew, T. W., Ubbens, J. R. & Stavness, I. Unsupervised domain adaptation for plant organ counting. In European Conference on
Computer Vision (eds Ayalew, T. W. et al.) 330–346 (Springer, 2020).
25. Gomes, D. P. S., & Zheng, L. Leaf Segmentation and Counting with Deep Learning: on Model Certainty, Test-Time Augmentation,
Trade-Offs. Preprint at arXiv:2012.11486. (2020).
26. Yang, K., Zhong, W. & Li, F. Leaf segmentation and classification with a complicated background using deep learning. Agronomy
10(11), 1721 (2020).
27. Kumar, J. P. & Domnic, S. Image based leaf segmentation and counting in rosette plants. Inform. Process. Agric. 6(2), 233–246
(2019).
28. Valente, J., & Giuffrida, M. V. Leaf counting from uncontrolled acquired images from greenhouse workers. Proceedings of the
Computer Vision Problems in Plant Phenotyping (CVPPP 2019), Long Beach, CA, USA, 17. (2019).
29. Jiang, B. et al. Leaf counting with multi-scale convolutional neural network features and fisher vector coding. Symmetry 11(4), 516
(2019).
30. Giuffrida, M. V., Doerner, P. & Tsaftaris, S. A. Pheno-deep counter: A unified and versatile deep learning architecture for leaf
counting. Plant J. 96(4), 880–890 (2018).
31. Ubbens, J., Cieslak, M., Prusinkiewicz, P. & Stavness, I. The use of plant models in deep learning: An application to leaf counting
in rosette plants. Plant Methods 14(1), 1–10 (2018).
32. Aich, S., & Stavness, I. Leaf counting with deep convolutional and deconvolutional networks. In Proceedings of the IEEE interna-
tional conference on computer vision workshops (pp. 2080-2089) (2017).
33. Kuznichov, D., Zvirin, A., Honen, Y., & Kimmel, R. Data augmentation for leaf segmentation and counting tasks in rosette plants.
In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops (pp. 0-0) (2019).
34. Itzhaky, Y., Farjon, G., Khoroshevsky, F., Shpigler, A., & Bar-Hillel, A. Leaf counting: Multiple scale regression and detection using
deep CNNs. In BMVC (p. 328) (2018).
35. Pape, J. M., & Klukas, C. Utilizing machine learning approaches to improve the prediction of leaf counts and individual leaf seg-
mentation of rosette plant images. Proceedings of the Computer Vision Problems in Plant Phenotyping (CVPPP), 1-12 (2015).
36. Klukas, C., Chen, D. & Pape, J. M. Integrated analysis platform: An open-source information system for high-throughput plant
phenotyping. Plant Physiol. 165(2), 506–518 (2014).
37. Minaee, S., Boykov, Y. Y., Porikli, F., Plaza, A. J., Kehtarnavaz, N., & Terzopoulos, D. Image segmentation using deep learning: A
survey. IEEE Transactions on Pattern Analysis and Machine Intelligence. (2021).
38. Qi, H., Zhang, Z., Xiao, B., Hu, H., Cheng, B., Wei, Y., Dai, J. Deformable convolutional networks coco detection and segmentation
challenge 2017 entry. ICCV COCO Challenge Workshop (2017).
39. Simonyan, K., & Zisserman, A. Very deep convolutional networks for large-scale image recognition. Preprint at arXiv:1409.1556.
(2014).
40. He, K., Zhang, X., Ren, S., & Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on
computer vision and pattern recognition (pp. 770-778) (2016).
41. Patel, V. & Mistree, K. A review on different image interpolation techniques for image enhancement. Int. J. Emerg. Technol. Adv.
Eng. 3(12), 129–133 (2013).
42. Khan, A., Sohail, A., Zahoora, U. & Qureshi, A. S. A survey of the recent architectures of deep convolutional neural networks.
Artif. Intell. Rev. 53(8), 5455–5516 (2020).
43. Rahman, M. A. & Wang, Y. Optimizing intersection-over-union in deep neural networks for image segmentation. In International
Symposium on Visual Computing (eds Rahman, M. A. & Wang, Y.) 234–244 (Springer, 2016).
44. Sudre, C. H., Li, W., Vercauteren, T., Ourselin, S., & Cardoso, M. J. Generalised dice overlap as a deep learning loss function for
highly unbalanced segmentations. In Deep learning in medical image analysis and multimodal learning for clinical (2017).
45. Hashim, F. A. & Hussien, A. G. Snake Optimizer: A novel meta-heuristic optimization algorithm. Knowl.-Based Syst. 242, 108320
(2022).
46. Hashim, F. A., Mostafa, R. R., Hussien, A. G., Mirjalili, S. & Sallam, K. M. Fick’s Law Algorithm: A physical law-based algorithm
for numerical optimization. Knowl.-Based Syst. 260, 110146 (2023).
47. Hu, G., Wang, J., Li, M., Hussien, A. G. & Abbas, M. EJS: Multi-strategy enhanced jellyfish search algorithm for engineering
applications. Mathematics 11(4), 851 (2023).
48. Hu, G., Zheng, Y., Abualigah, L. & Hussien, A. G. DETDO: An adaptive hybrid dandelion optimizer for engineering optimization.
Adv. Eng. Inform. 57, 102004 (2023).
49. Yu, H., Jia, H., Zhou, J. & Hussien, A. Enhanced Aquila optimizer algorithm for global optimization and constrained engineering
problems. Math. Biosci. Eng. 19(12), 14173–14211 (2022).
50. Sasmal, B., Hussien, A. G., Das, A., & Dhal, K. G. A Comprehensive Survey on Aquila Optimizer. Archives of Computational
Methods in Engineering, 1-28 (2023).
51. Izci, D., Ekinci, S. & Hussien, A. G. An elite approach to re-design Aquila optimizer for efficient AFR system control. PLoS ONE
18(9), e0291788 (2023).
52. Izci, D., Ekinci, S. & Hussien, A. G. Effective PID controller design using a novel hybrid algorithm for high order systems. PLoS
ONE 18(5), e0286060 (2023).
53. Hussien, A. G., Hashim, F. A., Qaddoura, R., Abualigah, L. & Pop, A. An enhanced evaporation rate water-cycle algorithm for
global optimization. Processes 10(11), 2254 (2022).
54. Chhabra, A., Hussien, A. G. & Hashim, F. A. Improved bald eagle search algorithm for global optimization and feature selection.
Alex. Eng. J. 68, 141–180 (2023).
Vol.:(0123456789)
www.nature.com/scientificreports/
55. Zheng, R. et al. A multi-strategy enhanced African vultures optimization algorithm for global optimization problems. J. Computat.
Des. Eng. 10(1), 329–356 (2023).
56. Hashim, F. A., Khurma, R. A., Albashish, D., Amin, M. & Hussien, A. G. Novel hybrid of AOA-BSA with double adaptive and
random spare for global optimization and engineering problems. Alex. Eng. J. 73, 543–577 (2023).
57. Hussien, A. G., Khurma, R. A., Alzaqebah, A., Amin, M., & Hashim, F. A. Novel memetic of beluga whale optimization with self-
adaptive exploration-exploitation balance for global optimization and engineering problems. Soft Computing, 1-39 (2023).
58. Hassan, M. H., Daqaq, F., Kamel, S., Hussien, A. G., & Zawbaa, H. M. An enhanced hunter-prey optimization for optimal power
flow with FACTS devices and wind power integration. IET Generation, Transmission & Distribution. (2023).
59. Ekinci, S. et al. Revolutionizing vehicle cruise control: An elite opposition-based pattern search mechanism augmented INFO
algorithm for enhanced controller design. Int. J. Computat. Intell. Syst. 16(1), 129 (2023).
60. Daqaq, F., Hassan, M. H., Kamel, S. & Hussien, A. G. A leader supply-demand-based optimization for large scale optimal power
flow problem considering renewable energy generations. Sci. Rep. 13(1), 14591 (2023).
61. Hassan, M. H., Kamel, S., Shaikh, M. S., Alquthami, T., & Hussien, A. G. Supply-demand optimizer for economic emission dispatch
incorporating price penalty factor and variable load demand levels. IET Generation, Transmission & Distribution. (2023).
62. Sasmal, B., Hussien, A. G., Das, A., Dhal, K. G., & Saha, R. Reptile search algorithm: Theory, variants, applications, and performance
evaluation. Archives of Computational Methods in Engineering, 1-29 (2023).
63. Elseify, M. A., Hashim, F. A., Hussien, A. G. & Kamel, S. Single and multi-objectives based on an improved golden jackal optimiza-
tion algorithm for simultaneous integration of multiple capacitors and multi-type DGs in distribution systems. Appl. Energy 353,
122054 (2024).
Author contributions
M.D.: Conceptualization, Methodology, Writing-original draft preparation. K.G.D.: Conceptualization, Visu-
alization, Investigation, Validation. A.D.: Conceptualization, Visualization, Investigation, Validation. A.G.H.:
Conceptualization, Visualization, Investigation, Validation. L.A.: Supervision, Validation, Writing-Reviewing
and Editing. A.G.: Software, Visualization, Writing-Reviewing.
Funding
Open access funding provided by Linköping University.
Competing interests
The authors declare no competing interests.
Additional information
Correspondence and requests for materials should be addressed to A.G.H.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Vol:.(1234567890)