Natural Mulches as an Option for Under-Plant Weed Administration in Mediterranean Flooded Grape Plantations: Impact on Agricultural Output
Received: 02-May-2023 / Manuscript No. jpgb-23-100811 / Editor assigned: 04-May-2023 / PreQC No. jpgb-23-100811 (PQ) / Reviewed: 18-May-2023 / QC No. jpgb-23-100811 / Revised: 23-May-2023 / Manuscript No. jpgb-23-100811 (R) / Published Date: 30-May-2023 DOI: 10.4172/jpgb.1000145
Abstract
For the complete implementation of site-specific weed management, which is currently a major challenge in modern agriculture, precise weed mapping is essential for sustainability, efficiency, and the maintenance of high crop yields and less chemically polluted agricultural lands. In this study, the robustness of the training epochs of the Convolutional Neural Network (CNN) model You Only Look Once (YOLO) v5s was evaluated for the creation of an automatic crop and weed classification using UAV images. The pictures were explained utilizing a jumping box and they were prepared on Google collaboratory north of 100, 300, 500, 600, 700, and 1000 ages. Sugarcane (Saccharum officinarum), banana trees (Musa), spinach (Spinacia oleracea), pepper (Capsicum), and weeds were all identified and categorized by the model. The model was trained over a number of epochs to find the best performance on the test set. When the test performance (classification accuracy, precision, and recall) started to drop, training was stopped. The result shows that the classifier's performance improved significantly as the number of training epochs increased, typically from 100 to 600. When the number of epochs was increased to 700, classification accuracy, weed precision, and recall were recorded at 65, 43, and 43%, respectively, compared to 67, 78, and 34% at 600 epochs, respectively. In the meantime, a slight decline was observed. When the epoch was increased to 1000, classification accuracy, weed precision, and recall of 65 percent, 45 percent, and 40 percent, respectively, were achieved, but this decline persisted. The findings revealed that the YOLOv5s training epoch has a significant impact on the model's robustness in automatic crop and weep classification, with 600 being the optimal epoch.
Introduction
One of the essential objectives of the Unified Countries is to destroy a wide range of yearning and unhealthiness by guaranteeing that everyone, especially youths and the people who are distraught people inside the general public appreciates admittance to a steady inventory of adequate and healthful food by 2030 [1]. Local farmers' livelihoods and skills must be improved, and they must be provided with an equitable supply of resources, such as land, technological advancements, and markets, in order to achieve sustainability in agricultural practices. There is evidence to suggest that the percentage of truly poor families worldwide decreases by 0.6% to 1.2% for every one percent increase in agricultural output. In the meantime, it was anticipated that population growth would reach 9.7 billion by 2050, requiring an increase in agricultural productivity of roughly 70% to meet the rising demand. Weeds, on the other hand, have a significant impact on crops, posing a serious threat to farms and reducing yields when not properly controlled and monitored. As a result, the achievement of the Sustainable Development Goals (SDGs) is primarily impacted by weeds, particularly in relation to agro-production and zero hunger [2]. The findings of Vilà et al. Non-native weeds may be responsible for 42 percent of crop production-related yield losses. These undesirable, noxious, and invasive plants prevent the growth of other crops, which has an effect on human activities, biological forces, and the economy of the nation.
However, prior to the development of hand tools for soil cultivation and weed eradication, human (hand) weeding, mechanized weeding, and herbicide sprays have been some of the most popular weed management methods. As a result, weed infiltration levels have been kept low and agricultural yields have increased worldwide as a result of these weed control and management strategies, but they are not without their share of drawbacks. The most significant issues with hand weeding are uneven weed control, rising labor costs, and decreasing labor supply [3]. Mechanical weed management, on the other hand, requires more soil turnover, which can alter the soil's morphology and reduce its nutrients. The effects after some time and the expense of mechanical weed control are not beneficial 100% of the time. Weeds that are resistant to herbicides, negative health effects, and pollution of the environment are just a few of the effects that chemical use has. In this way, it is critical to enhance existing contemporary weed administration procedures because of the hardships with respect to conventional weed control methodologies.
Pixel-based strategies are among the most often involved techniques in picture handling. The typical pixel-based approach ignores the potential for textural and spatial variables to improve algorithm accuracy because it focuses primarily on spectral data and performs computational evaluation at the pixel level [4]. However, Convolutional Neural Networks (CNNs) have recently evolved to take into account the textural, spatial, and spectral characteristics of photographs. This has improved or optimized classification, made them useful for managing a large amount of data simultaneously, and made the algorithm less ambiguous. Deep learning techniques are used in text synthesis, object classification, autonomous machine translation, and other fascinating applications. Precision agriculture has been used to anticipate the appearance of weeds and differentiate them from the crop zone [5]. This study presents the results of an evaluation of the performance of the deep learning algorithm You Only Look Once (YOLOv5s) in the creation of an automated weed classification scheme.
Literature review
Site Explicit Weed Administration (SSWM) is a methodology that includes changing weed control inside a homestead to consider varieties in weed populace size, dispersion, and variety. In farmlands, the number of inhabitants in weeds is regularly circulated sporadically. Thus, the underpinning of this center vital methodology is to offer a weed geological data map that will help the utilization of agrochemicals in a controlled framework to such an extent that the synthetics are applied straightforwardly to explicit requirements, while likewise utilizing different procedures, for example, consolidating any uses of plant subordinates which comprises of allelopathy impact, i.e., normal weed executioners to attempt to decrease substance tainting, and lessens soil, water, and air contamination.
The discovery and planning of weeds are the principal moves toward executing a SSWM strategy. This involves consolidating the sensor, handling strategies, and incitation of frameworks for weed map creation [6]. The temporal and spatial imaging resolution of conventional remote sensing technologies like piloting aircraft and earth observation satellites is lower, despite the fact that they can cover larger areas. Because of their inflated expense viability and convenience, automated ethereal vehicles (UAV) have exhibited an extraordinary possibility of working at lower elevations in this manner giving superior picture spatial goals in farming creation.
The weed management system will be completely automated, resulting in a significant reduction in the amount of human effort or input required to complete various tasks. In order to accomplish this, classification schemes that can distinguish weeds from crops have recently been proposed. Support Vector Machine, Random Forest (RF), Decision Tree, k-Nearest Neighbor (kNN), and Artificial Neural Networks are all examples of traditional object detection techniques that necessitate the labor of specialized specialists in order to locate relevant features (such as SIFT features, leaf digital features, and texture features). The procedure's flexibility must be altogether upgraded because of the muddled displaying process and testing execution [7]. Likewise, because of the way that the precision of the weed recognizable proof methodology vigorously depends on the classifier and preparing highlights used for the client characterized classes, especially with respect to the spatial and unearthly pixel thickness of the pictures, traditional strategies can't attempt the weed ID and order from crops ideally.
A few CNN-based designs have been proposed for object recognition which incorporates AlexNet, Quicker RCNN, GoogLeNet, LeNet-5, and single-stage calculation, for example, SSD, VGG, You Just Look Once (Consequences be damned), and so on. Region-based detection classifiers have been used by a number of researchers to identify objects. Using a CNN and a k-means model, for instance, weeds associated with soybean seedlings were identified in Liu & Chahl [8]. To upgrade the exactness of weed acknowledgment, further developed boundary values were gotten utilizing the k-implies preparing approach for uncontrolled markers. With an accuracy of 92.89 percent, the proposed method outperformed detection results using convolutional neural networks with random initialization and two-layer networks with no parameter tuning. Additionally, the authors demonstrated that altering the parameters has a significant effect on the improvement in accuracy. In fostering a weed recognition conspire, Lavania and Matey utilized picture division predicated on the 3D Otsu approach, to recognize ranch produce and weeds while performing characterization by packing 3D picture vectors utilizing Head Part Examination (PCA) strategy. Additionally, Urmashev et al. fostered a strategy for division and distinguishing proof of weeds commenced on Veil R-CNN. The ResNet-101 network was used to create a weeds-related semantic and geographic data extraction. The output modules performed the classification loss, regression, and segmentation calculations. The Mask R-CNN model outperformed the Sharp Mask and Deep Mask models with an average accuracy of 0.853. The idea of robotizing the course of weed expulsion utilizing AI calculations was explored by Sarvini et al. Four (4) assortments of monetary yields and two (2) assortments of weeds made up the information test that was procured and utilized. The developed classification models were compared to ANNs and CNNs in terms of their efficacy. The region-based identification classifiers' members have a high level of precision; however, the method requires a lot of processing and performs poorly in real time. dos Santos Ferreira et al. developed a method for distinguishing weeds in crops from those with broad, herbaceous leaves using an algorithm. Convolutional networks were utilized to give profoundly precise arrangement results to all weed types [9]. In any case, the grouping models used by the creators to classify weeds were not assessed either concerning velocity or execution.
Thusly, to give an answer for the difficulties of high computational intricacies, low continuous execution, and high handling time which are the significant holes seen in the explored writing, a solitary stage target identification calculation known as You Just Look Once (Consequences be damned) v5s was tested in this exploration. The exploration includes the obtaining of ethereal covering photos of the blended yield ranches utilizing a UAV fitted with a Red Green Blue (RGB) camera. These pictures are named, resized, and afterward used to prepare the Consequences be damned v5s model for the identification and characterization of the two yields and weeds.
Moreover, the examination involves investigating the effect of different preparation ages on the weed grouping plan's general exhibition exactness. Using full images in a single evaluation, the model divides a photograph into regions and predicts bounding boxes and possibilities for each region. Immediately, this exploration targets examining the exhibition of YOLOv5s in the programmed extraction and planning of weeds and yields in blended crop farmland. The performance was evaluated over six distinct training epochs (100, 300, 500, 600, 700, and 1000 epochs) in order to identify the epoch at which the model begins to become overfitted as it trains and to track the variations in accuracy as the number of epochs increased in order to select the epoch at which the automatic weed detection will be most accurate.
Execution design
The object localization and classification procedures are two steps in the object detection task. A bounding box is used by the object localization process to determine whether the object contains any objects of interest. The identified object is placed into one of the five classes that are used to classify the crops grown in the study area. The Consequences be damned v5 model works in a start to finish approach by doing both the article restriction and characterization straightforwardly. The backbone, neck, and head are its three components. The model's backbone extracts features from aerial photographs [10]. The Cross Stage Partial (CSP) served as the framework or backbone for quickly extracting characteristic features from the photographs, and the model's neck component was used to generate pyramids of the extracted features. The Way Conglomeration Organization (PANet) was carried out as the model's neck while the model's head plays out the last identification by applying anchor boxes to highlights like weeds and harvests. The implementation of quick training processes in the Darknet, an architecture that is compatible with GPU computations, is the primary contribution of YOLO v5. Additionally, the classifier detects even minute elements with remarkable precision. The structural plan of Just go for it v5 which portrays the ID technique of the yields and weeds is introduced [11].
Consequence of the programmed weed order
A 100-epoch evaluation of the weed automatic classification revealed classification accuracy, precision, and recall of 16 percent, 5 percent, and 13 percent, respectively. At 300 epochs, classification accuracy, precision, and recall were 65%, 46%, and 32%, respectively. At 500 epochs, classification accuracy, weed precision, and weed recall were 66%, 75%, and 27%, respectively. At 600 epochs, classification accuracy was 67%, while weed precision was 78%, and weed recall was 34%, respectively [12]. The precision of weeds decreased from 78% at 600 epochs to 43% at 700 epochs, while classification accuracy and weed recall of 65% and 43% were achieved, respectively, as the epochs increased to 700. Finally, classification accuracy, weed precision, and recall were found to be 65%, 45%, and 40% at 1000 epochs, respectively.
In addition, it was observed that the automatic weed classification classifier reached its highest level of precision (78 percent) at 600 epochs, and that accuracy began to decrease at 700 epochs, when it fell to (63 percent). In addition, increasing the number of epochs to 1000 had no significant effect on the accuracy of the weed classification [13].
Appendices 1 through 6 present the pattern of the weed visualization results, which demonstrates how weeds of varying sizes were identified within the mixed farm that included sugarcane, pepper, banana, and spinach. In Reference section 1 at 100 ages, the YOLOv5s model had the option to distinguish and group weeds inside the blended harvest ranch at around 63% in accuracy. This was because of the way that the age utilized was excessively little for the model to advance totally. When the model was at 300 epochs, as shown in Appendix 2, the algorithm was only able to classify the weeds with a 63% accuracy. At 500 epochs, the model correctly identified the weeds, as shown in Appendix 3. At 600 ages as displayed in Supplement 4, the weeds were ordered roughly to an accuracy of 78% inside the farmland [14]. In addition, at 700 epochs in Appendix 5, weed class was identified and classified with a precision of approximately 63%, and at 1000 epochs in Appendix 6, weeds were classified with a precision of 51% [15]. From the outcome acquired from the various ages, it was found that at 600 ages, the accuracy of weed arrived at its ideal condition of 78% and started to decline when the scope of ages expanded from 600 ages to 700 ages at 63%, and afterward to 51% at 1000 ages. In addition, it was discovered that after 700 epochs, there was no significant improvement.
Conclusion
Both automatic weed recognition and highly accurate crop type classification in a mixed crop farmland have been demonstrated to be very robust applications of deep learning algorithms, and CNNs in particular have the potential to improve the efficiency of real-time weed management. In this review, Just go for it v5s, a CNN model has been executed for crop grouping and weed recognition, while likewise assessing the model's presentation over various preparation ages. Loss function graphs, precision and recall graphs, confusion matrix, and additional validation metrics like the F1 score, accuracy, recall, and precision were used to assess the developed model's accuracy. The study found that as the number of training epochs increased, the crop and weed classification scheme got better over time, even though the model reached its best performance after 600 epochs, became saturated, and started to decline.
Likewise, the preparation misfortunes diminished all through the preparation interaction which suggested that the model was learning and in the long run leveled out at 700 ages where the weed accuracy started to fundamentally drop. In conclusion, when identifying and classifying weeds from mixed-crop farmland, the YOLOv5s model achieved comparable detection accuracies while exhibiting a significant advantage in computation times. Since around 254 picture matches were utilized for this trial and error, future exploration endeavors will consider examining the exhibition of the model with more weed endlessly pictures with better spatial goal.
Acknowledgement
None
Conflict of Interest
None
References
- Gil JBD, Reidsma P, Giller K, Todman L, Whitmore A, et al. (2019) . Ambio 48: 685-698.
- Baillo EH, Kimotho RN, Zhang Z, Xu P (2020) . Genes (Basel) 10: 771.
- Ng DWK, Abeysinghe JK, Kamali M (2018) Agriculture in 2050: recalibrating targets for sustainable intensification. Int J Mol Sci 19: 3737.
- Paul S, Datta BK, Ratnaparkhe MB, Dholakia BB (2022) . Mol Biotechnol 64: 221-244.
- Leke WN, Brown JK, Ligthart MK, Sattar N, Njualem DK (2012) . Virus Res 163: 229-37.
- Plaisier CL, Lo FY, Ashworth J, Brooks AN, Beer KD, et al. (2014) . BMC Syst Biol 8: 122.
- Walsh MJ, Broster JC, Lazaro LMS, Norsworthy JK et al. (2018) . Pest Manag Sci 74: 2235-2245.
- Anten NPR, Chen BJW (2021) . Plant Cell Environ 44: 3475-3478.
- Gerhards R, Schappert A (2020) . Pest Manag Sci 76: 42-46.
- Ruiz MP, Carballido J, Agüera J, Lizana AR (2013) . Sensors (Basel) 13: 3313-30.
- Gianessi LP (2013) . Pest Manag Sci 69: 1099-1105.
- Kumar R, Choudhary JS, Mishra JS, Mondal S, Poonia S, et al. (2022) . Sci Rep 12: 3753.
- Ghosh D, Brahmachari K, Skalický M, Roy D, Das A, et al. (2022) . PLoS One 17: e0262586.
- Annett R, Habibi HR, Hontela A (2014) . J Appl Toxicol 34: 458-479.
- Hoppin JA (2014) Occup Environ Med 71: 80.
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
, ,
Citation: Xn S (2023) Natural Mulches as an Option for Under-Plant Weed Administration in Mediterranean Flooded Grape Plantations: Impact on Agricultural Output. J Plant Genet Breed 7: 145. DOI: 10.4172/jpgb.1000145
Copyright: © 2023 Xn S. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Share This Article
黑料网 Journals
Article Tools
Article Usage
- Total views: 1097
- [From(publication date): 0-2023 - Nov 22, 2024]
- Breakdown by view type
- HTML page views: 1020
- PDF downloads: 77