WO2024009284A1 - Systems and methods to identify plant species and varieties using computer vision and intelligent technologies - Google Patents
Systems and methods to identify plant species and varieties using computer vision and intelligent technologies Download PDFInfo
- Publication number
- WO2024009284A1 WO2024009284A1 PCT/IB2023/057044 IB2023057044W WO2024009284A1 WO 2024009284 A1 WO2024009284 A1 WO 2024009284A1 IB 2023057044 W IB2023057044 W IB 2023057044W WO 2024009284 A1 WO2024009284 A1 WO 2024009284A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- images
- geographic areas
- training
- image
- patches
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/17—Terrestrial scenes taken from planes or by drones
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/188—Vegetation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/194—Terrestrial scenes using hyperspectral data, i.e. more or other wavelengths than RGB
Definitions
- the present disclosure relates to systems and methods for classifying agricultural plant varieties and species using computer vision and machine learning classifiers.
- sugarcane variety classification currently are achieved in two different ways: analyzing genetic information extracted from live individuals looking for specific patterns manifested in different species, or with a protocol with more than 8 steps to visually classify the varieties based on morphological characteristics such as form, texture and color of the leaves, nodes size, and internode aspects, among others.
- biometric analysis is considered important as depending on field and climatic conditions during production seasons, some morphological characteristics should be impacted and demonstrate relevant phenotypic alterations.
- the first method (analyzing genetic information from extracted plants) is costly and depends on physically collecting samples of each crop field in order to classify the species or varieties. Additionally, for some species, as for example sugarcane, the complexity of the polyploid and aneuploid genome of modem hybrid varieties, imposes challenges on development of variety classification methods base on genetic information. [0006]
- the second method is time consuming, subject to variation due to environment conditions and highly dependent on a very skilled personnel doing a detailed visual inspection for the classification, especially to identify color variations and formats, either in person or with high quality images capturing details of the plant being analyzed from specific angles (e.g., details from leaves, stalks, and internodes, among others). Thus, both processes do not scale up for large areas such as the whole country (for instance sugar cane being grown in the entire country of Brazil, as well large area of India, China, Thailand and other countries).
- Physiological distinction between plant species and varieties affects, e.g.., carbon and mineral cycles and methane emissions, interfering on conclusions about the environmental impacts of land use.
- Examples of important crops in terms of cultivated area around the world are: barley, cassava, cotton, groundnuts or peanuts, maize, millet, oil palm fruit, potatoes, rapeseed or canola, rice, rye, sorghum, soybeans, sugar cane, sugar beets, sunflower, citrus, and wheat.
- a method for identifying plants growing in one or more geographic areas comprises: receiving a plurality of images of the one or more geographic areas that contain one or more plants, selecting one or more images from the plurality of images for classification, cropping the selected images based on one or more shapes that delimit one or more boundaries of the one or more geographic areas, converting the selected images into one or more image patches, wherein each patch of the one or more patches represents a subset of a geographic area of the one or more geographic areas, applying a machine learning classifier to each patch of the one or more image patches to generate one or more classifications, wherein the machine learning classifier is created using a supervised training process that comprises using one or more annotated images to train the machine learning classifier, and identifying one or more plants growing in the one or more geographic areas based on the one or more classifications generated by the machine learning classifier.
- the supervised training process comprises: receiving a plurality of training images from one or more geographic areas that contain a known plant, cropping the training images based on one or more shapes that delimit one or more boundaries of the one or more geographic areas that contain a known plant, converting the training images into one or more training image patches, wherein each patch of the one or more patches represents a subset of a geographic area of the one or more geographic areas that contain a known plant, applying one or more annotations to the training image patches to indicate a plant contained within each training image patch, and processing each training image and its corresponding one or more annotations by the machine classifier.
- the annotations of the training images can identify a specific target (e.g., a variety or a species) or can identify a target as pertaining to a specific group (e.g., a group of varieties or species).
- the model can classify a target or a group of targets by binary or multiclass classifications (using one or multiple steps).
- the received plurality of training images from one or more geographic areas are captured from a plurality of months during a plantation lifecycle.
- the method comprises applying a segmentation process to each of the selected one or more images.
- the segmentation process comprises calculating a normalized difference vegetation index (ND VI) for each image of the selected one or more images.
- ND VI normalized difference vegetation index
- the segmentation process comprises calculating a soil-adjusted vegetation index (SAVI) for each image of the selected one or more images.
- SAVI soil-adjusted vegetation index
- identifying the one or more plants growing in the one or more geographic areas based one or more classifications generated by the machine learning classifier comprises receiving a classification of each image patch of the one or more image patches and determining a majority classification from the received classification of each image patch.
- identifying the one or more plants growing in the one or more geographic areas based on the one or more classifications generated by the machine learning classifier comprises classifying the plant growing in the one or more geographic areas based on the majority classification corresponding to the one or more image patches corresponding to the geographic area.
- the machine classifier is implemented using one or more convolution neural networks (CNNs). Additionally or alternatively, recurrent neural networks and/or classical statistical models for image and time series processing may be used.
- the CNN includes a plurality of convolutional blocks selected from the group consisting of a batch normalization block, a dropout block, and ReLU activation block.
- the received plurality of images are multispectral images.
- the received plurality of images are hyperspectral images.
- the received plurality of images are RGB images. Additionally or alternatively, the received plurality of images are SAR (Synthetic Aperture Radar) images.
- the received plurality of images of the one or more geographic areas are taken from a satellite.
- the received plurality of images of the one or more geographic areas are taken from an airplane.
- the received plurality of images of the one or more geographic areas are taken from an unmanned aerial vehicle (UAV).
- UAV unmanned aerial vehicle
- converting the selected images into one or more patches comprises up- sampling the one or more image patches.
- a system for identifying plants growing in one or more geographic areas comprises: one or more processors, memory comprising a local storage, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for: receiving a plurality of images of the one or more geographic areas that contain one or more plants, selecting one or more images from the plurality of images for classification, cropping the selected images based on one or more shapes that delimit one or more boundaries of the one or more geographic areas, converting the selected images into one or more image patches, wherein each patch of the one or more patches represents a subset of a geographic area of the one or more geographic areas, applying a machine learning classifier to each patch of the one or more image patches to generate one or more classifications, wherein the machine learning classifier is created using a supervised training process that comprises
- the supervised training process comprises: receiving a plurality of training images from one or more geographic areas that contain a known plant, cropping the training images based on one or more shapes that delimit one or more boundaries of the one or more geographic areas that contain a known plant, converting the training images into one or more training image patches, wherein each patch of the one or more patches represents a subset of a geographic area of the one or more geographic areas that contain a known plant, applying one or more annotations to the training image patches to indicate a plant contained within each training image patch, and processing each training image and its corresponding one or more annotations by the machine classifier.
- the annotations of the training images can identify a specific target (e.g., a variety or a species) or can identify a target as pertaining to a specific group (e.g., a group of varieties or species).
- the model can classify a target or a group of targets by binary or multiclass classifications (using one or multiple steps).
- the received plurality of training images from one or more geographic areas are captured from a plurality of months during a plantation lifecycle.
- the segmentation process comprises calculating a normalized difference vegetation index (ND VI) for each image of the selected one or more images.
- ND VI normalized difference vegetation index
- the segmentation process comprises calculating a soil-adjusted vegetation index (SAVI) for each image of the selected one or more images.
- SAVI soil-adjusted vegetation index
- the received plurality of images are multispectral images.
- the received plurality of images are hyperspectral images.
- the received plurality of images are RGB images. Additionally or alternatively, the received plurality of images are SAR (Synthetic Aperture Radar) images.
- the supervised training process comprises: receiving a plurality of training images from one or more geographic areas that contain a known plant, cropping the training images based on one or more shapes that delimit one or more boundaries of the one or more geographic areas that contain a known plant, converting the training images into one or more training image patches, wherein each patch of the one or more patches represents a subset of a geographic area of the one or more geographic areas that contain a known plant, applying one or more annotations to the training image patches to indicate a plant contained within each training image patch, and processing each training image and its corresponding one or more annotations by the machine classifier.
- the annotations of the training images can identify a specific target (e.g., a variety or a species) or can identify a target as pertaining to a specific group (e.g., a group of varieties or species).
- the model can classify a target or a group of targets by binary or multiclass classifications (using one or multiple steps).
- the received plurality of training images from one or more geographic areas are captured from a plurality of months during a plantation lifecycle.
- the one or more processors are caused to apply a segmentation process to each of the selected one or more images.
- the segmentation process comprises calculating a normalized difference vegetation index (ND VI) for each image of the selected one or more images.
- ND VI normalized difference vegetation index
- the segmentation process comprises calculating a soil-adjusted vegetation index (SAVI) for each image of the selected one or more images.
- SAVI soil-adjusted vegetation index
- identifying the one or more plants growing in the one or more geographic areas based on the one or more classifications generated by the machine learning classifier comprises classifying the plant growing in the one or more geographic areas based on the majority classification corresponding to the one or more image patches corresponding to the geographic area.
- the CNN includes a plurality of convolutional blocks selected from the group consisting of a batch normalization block, a dropout block, and ReLU activation block.
- the received plurality of images are multispectral images.
- the received plurality of images are hyperspectral images.
- the received plurality of images are RGB images. Additionally or alternatively, the received plurality of images are SAR (Synthetic Aperture Radar) images.
- the received plurality of images of the one or more geographic areas are taken from an airplane.
- the received plurality of images of the one or more geographic areas are taken from an unmanned aerial vehicle (UAV).
- UAV unmanned aerial vehicle
- FIG. 1 illustrates an exemplary process for training and validating a machine learning classifier according to examples of the disclosure.
- FIG. 3 illustrates an exemplary process for classifying unknown crop species and varieties using a machine classifier according to examples of the disclosure.
- FIG. 4 illustrates a visual depiction of the image processing steps used to classify an unknown crop species and varieties using a machine learning classifier according to examples of the disclosure.
- FIG.5 illustrates an example of a computing device according to examples of the disclosure.
- Certain aspects of the present invention include process steps and instructions described herein in the form of an algorithm. It should be noted that the process steps and instructions of the present invention could be embodied in software, firmware, or hardware, and, when embodied in software, could be downloaded to reside on and be operated from different platforms used by a variety of operating systems.
- the present invention also relates to a device for performing the operations herein.
- This device may be specially constructed for the required purposes, or it may comprise a generalpurpose computer selectively activated or reconfigured by a computer program stored in the computer.
- a computer program may be stored in a non-transitory, computer-readable storage medium, such as, but not limited to, any type of disk, including floppy disks, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application-specific integrated circuits (ASICs), or any type of media suitable for storing electronic instructions and each coupled to a computer system bus.
- any type of disk including floppy disks, optical disks, CD-ROMs, magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, application-specific integrated circuits (ASICs), or any type of media suitable for
- the computers referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.
- the processing frameworks referred to herein can include a cloud architecture, with process orchestration systems for on-demand processing scalability. Any of the processing operations described herein may be performed by a single processor, multiple processors, distributed processor systems, and/or cloud architecture processing systems.
- the methods, devices, and systems described herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may also be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the required method steps. The required structure for a variety of these systems will appear from the description below.
- the present invention is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein.
- Identifying the species, sub-species or the variety of a single plant can be a simple and quick process.
- One method can be to physically extract a sample of the plant growing in a field and performing a genetic analysis to ascertain what species of plant or variety of that species it is.
- Another method can involve performing a visual inspection of the plant and looking for physical markers (such as leave, stalks, internodes, etc.) that indicate the species or variety of plant. While the above methods can be straightforward for a single plant, these methods may not be efficient when attempting to analyze a large area or field of plants. For instance, it may not be efficient to extract physical samples of every plant growing in a field, and visual identification may prove to be too cumbersome and time intensive to be performed at scale on a large field.
- the methodologies available for identification require a long-term process and a very skilled person to yield accurate results.
- What is needed is a method for classifying plants that can provide quick and accurate results, and can be used for classifications of plant species and varieties in large geographic areas.
- employing the use of machine-learning based classifiers to automatically and efficiently classify the plant species and varieties growing in a large geographic area can allow for the classification of plant species and/or varieties massively and quickly in very large territories.
- the process of processing images used to train and apply a classifier can be tailored to ensure that the classification generated by such classifiers is accurate to an acceptable level.
- the method for building and training the classifier can be important to ensure that the classifier makes classifications with a high-level of accuracy.
- the method of using the classifier can also be important to ensure a high-fidelity classification.
- FIG. 1 illustrates an exemplary process for training and validating a machine learning classifier to classify plant species and varieties according to examples of the disclosure.
- the process 100 of FIG. 1 can be used to generate an annotated data set that can be used to then train a machine learning classifier to automatically and efficiently classify plant species and varieties found in a particular geographic area.
- the process 100 can begin at step 102 wherein images are selected to train and validate the classifier.
- the images used to train and validate a classifier can be selected from one or more commercially available sources such as the Sentinel-2 Satellite or other image databases.
- more than one image per area can be used to train and validate a classifier, adding variation to the computer method and system.
- distinct classifiers can be trained for each month during the plantation lifecycle and the multiple classifiers can be combined in the end of the process, which can result in larger plantations areas identified with increased confidence and accuracy.
- the Sentinel-2 the Military Grid Reference System (MGRS) can be used as reference to download a tile image that covers an area of approximately 100km x 100km (1 million hectares).
- MGRS Military Grid Reference System
- Each Sentinel-2 image has 13 spectral bands with different wavelengths, from the Visible (VNIR) and Near Infra-Red (NIR) to the Short Wave Infra- Red (SWIR).
- VNIR Visible
- NIR Near Infra-Red
- SWIR Short Wave Infra- Red
- the metadata of the satellite images can be analyzed to find available images in the period that corresponds to the beginning of the harvest season.
- images corresponding to the beginning of the harvest season for instance the month of April in Brazil
- two or more images from each area can be selected during the crop lifecycle.
- two or more images from each area can be selected during the period that corresponds to the harvest season of the crop.
- at least twelve images from the same location are selected during crop lifecycle. Images from the same location but different months or periods can be used to train and validate classifiers that will be able to learn specific characteristics of each species or variety that are dominant in that period, adding to the model morphological and environmental condition variations captured by satellite images, which can help with the accurate classification of different species and varieties throughout the year.
- the combination of the results obtained from the classification in the different months or periods can be combined in order to (i) improve the accuracy of the classification for different species and varieties, (ii) be able to classify early stage crop fields that might be too young in some images, as well as crop fields that might be covered by clouds in multiple images captured by the satellite, and (iii) increase the confidence and accuracy of the classification based on the consistence in the results throughout multiple months or periods for each crop field, (e.g., how many months provides the same classification, representing consensus from the different trained classifiers for each month or period.) Additionally or alternatively, a soil and/or vegetation indexes can be used to identify and discard images from areas with low vegetation cover.
- SAVI Soil-adjusted vegetation index
- ND VI Normalized difference vegetation index
- SAVI index lower than 0.3 are used to identify images that should be disregarded.
- the process 100 can move to step 104 wherein the annotations applied to the image set for classifier training can be generated.
- the labeled dataset for training and validating the model can be created using an annotated geospatial database that includes information of crop fields in which the different species and varieties are planted.
- a large annotated database can be critical to properly train the model, for instance, in one or more examples; in order to generate the necessary annotations for a training set to identify sugar cane varieties a proprietary database comprising 2.5 million hectares of sugar cane crop fields with different varieties planted can be utilized to generate the needed annotations.
- annotations of training images can identify a specific target (e.g., a variety or a species) or can identify a target as pertaining to a specific group (e.g., a group of varieties or species).
- the model can classify a target or a group of targets by binary or multiclass classifications (using one or multiple steps).
- each image selected at step 102 can be annotated with the information extracted from the proprietary database at step 104.
- step 102 and 104 occur concomitantly (in parallel), at the same processing time.
- each image is cropped and sliced to generate individual samples used for training the classifier.
- the shapes that delimit each crop field containing a specific species and variety can be used to crop the image of that sugar cane plantation.
- the shape of crop fields is defined based on information from a pre-existent proprietary georeferenced database.
- each identified crop field can be further sliced in patches of 8 pixels by 8 pixels (representing an area of 80m by 80m), finally creating the labeled dataset that are going to be used for training the model.
- a fixed size patch can be used to allow for the proper training of a Neural Network (e.g., to feed the input layer).
- different patch sizes can be empirically determined in order to define the optimal size for the specific images selected to train the classifier and associated with a particular plant species or variety to be classified (e.g., image resolution and spectral bands available, amount of labeled data, shapes of the crop fields).
- the step of cropping the image into fixed sized patches can include ignoring the pixels at the border of each crop field, through an erosion operation, to avoid the noise that might be present due to the satellite sensors receiving the reflectance both from the plantation as well as neighboring areas, which can contain naked soil, other plantations and plantations from other species, a street, a construction or any other material.
- the slicing process can also include using a sliding window technique to perform up-sampling in order to reduce the challenges due to the limited area available for the species/variety that have smaller market share.
- the process of using a sliding window in the cropping process can include cropping regions with partial overlap, and moving the crop mask in smaller steps than the patch size.
- moving the mask in steps of four pixels instead of eight pixels can make the required the number of samples obtained for training roughly (depending on shape boundaries) multiplied by 4. Up- sampling can be critical to achieving a reasonable number of samples for each species or variety, achieving a balanced training set, and allowing the model to learn how to distinguish between different species or varieties.
- the process 100 can move to step 108, wherein the annotations generated at step 104 can be applied to each individual slice created at step 106. Furthermore, once the annotations have been applied at step 108, the process 100 can move to step 110 wherein each generated slice can be normalized to better handle differences in the reflectance captured during the satellite imagining process, thus making sure that the comparison between samples is more accurate.
- the normalization of the sample images at step 110 can be done by adjusting the different ranges in the reflectance of each band to the same range with a linear transformation using the min/max values of each band.
- different values can be chosen as min/max to remove the effect of outliers in the normalization and/or to saturate the image in certain ranges.
- a transformation based on the distribution or histogram of each band can be employed to normalize the images in different ways for each bucket or percentile.
- a robust scaler transform can be used to normalize the image scaling features using statistics that are robust to outliers.
- the process 100 can move to step 112 wherein the labeled dataset can be split to create training, validation, and testing sets.
- the slices from a single plantation can be kept together in the same set to avoid information leakage between training, validation, and testing sets.
- the training set must be balanced, e.g., with the same number of samples for each species/ variety, thus restricting the number of samples that can be used by the species/variety that have the least number of known crop fields in a particular geospatial labeled database.
- the process 100 at step 112 can include down- sampling, e.g., randomly choosing the ones to be used during the training process, while for the species/varieties that have a lower number of samples, the process 100 can employ an up-sampling technique using sliding windows as described above.
- the process 100 can move to step 114 wherein the classifier is trained using the training set.
- the training set can be fed to a machine learning model, so the machine learning model can learn how to classify the species or varieties of each image patch of plantations.
- the training set can be fed to a neural network model, so the neural network can learn how to classify the species or varieties of each image patch of plantations.
- a deep neural network algorithm was chosen to avoid the need of building hand-engineered features that requires prior knowledge and human effort in feature design, given that experiments with traditional models using commonly used features such as average and standard deviation of the reflectance for each spectral bands and indexes such as ND VI (Normalized Difference Vegetation Index) and others didn’t produce accurate results.
- a Convolutional Neural Network a class of deep neural networks, can be utilized as the classifier due to its distinguishing features towards image classification problems, allowing the exploitation of strong spatial local correlation present in images, translational invariance, and shared- weights architecture, which can reduce the number of parameters learned and memory requirements for training such networks, allowing CNNs to better generalize on vision oriented tasks.
- recurrent neural networks, other artificial neural networks, other machine learning architectures, and/or classical statistical models for image and time series processing may be used.
- a CNN architecture can be designed using three convolutional blocks, each one including convolution, batch normalization, dropout, and ReLU activation, followed by a fully connected layer.
- the input to the CNN can be an 8x8x13 multispectral image (8 pixels by 8 pixels by 13 spectral channels), and the output can be the predicted plant species or variety code.
- the size of the patches, described throughout this disclosure as 8x8x13, is meant only as an example, and the size can be adjusted depending on the resolution of the image source, the number of spectral channels available in the image, the average size of crop fields and the amount of labeled fields available.
- a CNN architecture can be used to allow the network to be deep enough to abstract the needed knowledge representations to learn how to identify the different plant species or variety, while controlling the depth to keep the network feasible to be trained with the amount of labeled data and image samples we have available.
- the CNN can be trained using Adam and a decaying learning rate to make the model explore more in the beginning and perform fine adjustments in the end.
- This CNN algorithm can be chosen due to its adaptive learning rate optimization that allows huge performance gains in terms of speed of training deep neural networks.
- the number of convolutional blocks in the neural network and other hyperparameters such as the learning rate can also be adjusted according to the dataset available for training and applying the model, using experimentation and hyperparameter optimization techniques.
- FIG. 2 illustrates a visual depiction of the image processing steps used to train a machine learning classifier according to examples of the disclosure.
- the process 200 of FIG. 2 can represent a visual depiction of the process 100 described above with respect to FIG. 1.
- the images from an image database can be cropped as described above with respect to step 106.
- the shapes that delimit each crop field containing a specific species/variety can be further used to crop the image of that plantation as depicted at step 204.
- the process can move to step 206 wherein the 8x8 patches are created as described above with respect to step 106 of FIG. 1.
- each individual annotated patch can be fed into a CNN so as to train the classifier at step 208.
- the output of the CNN can produce one or more classifications for each patch fed into the CNN.
- the trained model can be used to predict the species or variety of an unknown plantation.
- applying the classifier can involve the same process as training the classifier with respect to cropping, slicing, and preparing images for input into the classifier.
- two additional steps can be included in the process when applying the classifier to classify an unknown crop species or variety: a segmentation step in the beginning of the process, to segment each crop field unit in a given region of interest, and a voting step, to combine the results of each 8x8 patch from the same crop unit into a more robust prediction.
- FIG. 3 illustrates an exemplary process for classifying unknown crop species/variety using a machine classifier according to examples of the disclosure.
- the process 300 of FIG. 3 wherein similar to step 102 of FIG. 1, images for classification are selected. Similar to step 102, the images selected at step 302 can be selected based on the growing season in which the image was captured or other features that ensure that the crops to be classified are at their peak visibility in the images.
- the process 300 can move to step 304, wherein the selected images undergo a segmentation process.
- the image segmentation process (carried out at step 304) can involve calculating the ND VI (Normalized Difference Vegetation Index) on the regions of interest to visually enhance the naked soil that usually separates different fields, and applying the Felzenszwalb graph-based segmentation algorithm (Felzenszwalb & Huttenlocher, 2004) on the resulting image to segment the individual crop fields.
- ND VI Normalized Difference Vegetation Index
- Other indexes in replacement of ND VI such as SAVI - Soil- Adjusted Vegetation Index
- other segmentation algorithms in replacement of Felzenszwalb could also be used with no detriment to the invention disclosed herein.
- a single field or region can comprise multiple slices. Therefore, to ensure accurate classification, once the slices have each been classified at step 308, the process 300 can move to step 310 wherein a voting process is applied to the classified slices in order to ensure maximum accuracy of the classification results.
- the voting process performed at step 310 can combine the results of the model applied to each patch (e.g., slice) from an individual field, gathering consensus (majority voting) and improving the accuracy of the final prediction, choosing the predominant classification as the species/variety for that crop field.
- FIG. 4 illustrates a visual depiction of the image processing steps used to classify an unknown crop species/varieties using a machine learning classifier according to examples of the disclosure.
- the process depicted in FIG. 4 shows that the cropped images (cropped at step 306) and the field segmentation algorithm (applied at step 304) can be used to create one or more crop units 406. After the crop units (which represent a particular field of the same crop) are created, they are sliced into 8x8 patches as depicted at step 408. Each patch is then fed into a CNN classifier to perform the classification as depicted at 410.
- the output of the classifier can produce a plurality of classified patches in which each patch fed into the CNN can be assigned a classification as depicted at 412. As depicted at 414, using the voting procedure described in at step 310 with respect to FIG. 3, each of the crop units generated at 406 can be classified. Finally, in one or more examples, the classified units can be stored and used to create a visualization (representing the classifier results) as depicted at step 418.
- FIG. 5 illustrates an example of a computing device in accordance with one embodiment.
- Device 500 can be a host computer connected to a network.
- Device 500 can be a client computer or a server.
- device 500 can be any suitable type of microprocessor-based device, such as a personal computer, workstation, server, or handheld computing device (portable electronic device), such as a phone or tablet.
- the device can include, for example, one or more of processors 502, input device 506, output device 508, storage 510, and communication device 504.
- Input device 506 and output device 508 can generally correspond to those described above and can either be connectable or integrated with the computer.
- Input device 506 can be any suitable device that provides input, such as a touch screen, keyboard or keypad, mouse, or voice-recognition device.
- Output device 508 can be any suitable device that provides output, such as a touch screen, haptics device, or speaker.
- Software 512 which can be stored in storage 510 and executed by processor 502, can include, for example, the programming that embodies the functionality of the present disclosure (e.g., as embodied in the devices as described above).
- Software 512 can also be stored and/or transported within any non-transitory computer-readable storage medium for use by or in connection with an instruction execution system, apparatus, or device, such as those described above, that can fetch instructions associated with the software from the instruction execution system, apparatus, or device and execute the instructions.
- a computer-readable storage medium can be any medium, such as storage 510, that can contain or store programming for use by or in connection with an instruction execution system, apparatus, or device.
- Software 512 can also be propagated within any transport medium for use by or in connection with an instruction execution system, apparatus, or device, such as those described above, that can fetch instructions associated with the software from the instruction execution system, apparatus, or device and execute the instructions.
- a transport medium can be any medium that can communicate, propagate, or transport programming for use by or in connection with an instruction execution system, apparatus, or device.
- the transport readable medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, or infrared wired or wireless propagation medium.
- Device 500 may be connected to a network, which can be any suitable type of interconnected communication system.
- the network can implement any suitable communications protocol and can be secured by any suitable security protocol.
- the network can comprise network links of any suitable arrangement that can implement the transmission and reception of network signals, such as wireless network connections, T1 or T3 lines, cable networks, DSL, or telephone lines.
- Device 500 can implement any operating system suitable for operating on the network.
- Software 512 can be written in any suitable programming language, such as C, C++, Java, or Python.
- application software embodying the functionality of the present disclosure can be deployed in different configurations, such as in a client/server arrangement or through a Web browser as a Web-based application or Web service, for example.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Remote Sensing (AREA)
- Astronomy & Astrophysics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202263359622P | 2022-07-08 | 2022-07-08 | |
| US63/359,622 | 2022-07-08 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2024009284A1 true WO2024009284A1 (en) | 2024-01-11 |
Family
ID=87557538
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/IB2023/057044 Ceased WO2024009284A1 (en) | 2022-07-08 | 2023-07-07 | Systems and methods to identify plant species and varieties using computer vision and intelligent technologies |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2024009284A1 (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119107553A (en) * | 2024-09-12 | 2024-12-10 | 广东省国土资源测绘院 | Characteristic crop identification method and system based on satellite-borne and airborne hyperspectral remote sensing |
| TWI895113B (en) * | 2024-09-20 | 2025-08-21 | 瑞竣科技股份有限公司 | Vegetation satellite image identification system and method |
| CN120875185A (en) * | 2025-09-26 | 2025-10-31 | 西安科技大学 | Dynamic monitoring method for vegetation coverage of photovoltaic area based on artificial intelligence |
Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10852421B1 (en) * | 2019-01-24 | 2020-12-01 | Descartes Labs, Inc. | Sparse phase unwrapping |
-
2023
- 2023-07-07 WO PCT/IB2023/057044 patent/WO2024009284A1/en not_active Ceased
Patent Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10852421B1 (en) * | 2019-01-24 | 2020-12-01 | Descartes Labs, Inc. | Sparse phase unwrapping |
Non-Patent Citations (1)
| Title |
|---|
| SANDEEP KUMAR M ET AL: "Delineation of field boundary from multispectral satellite images through U-Net segmentation and template matching", ECOLOGICAL INFORMATICS, ELSEVIER, AMSTERDAM, NL, vol. 64, 18 July 2021 (2021-07-18), XP086757844, ISSN: 1574-9541, [retrieved on 20210718], DOI: 10.1016/J.ECOINF.2021.101370 * |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119107553A (en) * | 2024-09-12 | 2024-12-10 | 广东省国土资源测绘院 | Characteristic crop identification method and system based on satellite-borne and airborne hyperspectral remote sensing |
| TWI895113B (en) * | 2024-09-20 | 2025-08-21 | 瑞竣科技股份有限公司 | Vegetation satellite image identification system and method |
| CN120875185A (en) * | 2025-09-26 | 2025-10-31 | 西安科技大学 | Dynamic monitoring method for vegetation coverage of photovoltaic area based on artificial intelligence |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Ninomiya | High-throughput field crop phenotyping: current status and challenges | |
| Palei et al. | A systematic review of citrus disease perceptions and fruit grading using machine vision | |
| Moghimi et al. | Aerial hyperspectral imagery and deep neural networks for high-throughput yield phenotyping in wheat | |
| Xu et al. | Instance segmentation method for weed detection using UAV imagery in soybean fields | |
| Pathak et al. | A review of unmanned aerial vehicle-based methods for plant stand count evaluation in row crops | |
| Bakhshipour et al. | Weed segmentation using texture features extracted from wavelet sub-images | |
| Abdullahi et al. | Convolution neural network in precision agriculture for plant image recognition and classification | |
| WO2024009284A1 (en) | Systems and methods to identify plant species and varieties using computer vision and intelligent technologies | |
| Moeinizade et al. | An applied deep learning approach for estimating soybean relative maturity from UAV imagery to aid plant breeding decisions | |
| Tong et al. | Deep Learning for Image-Based Plant Growth Monitoring: A Review. | |
| CN113657158B (en) | Google EARTH ENGINE-based large-scale soybean planting area extraction algorithm | |
| Safari et al. | A review on automated detection and assessment of fruit damage using machine learning | |
| Mumtaz et al. | Integrated digital image processing techniques and deep learning approaches for wheat stripe rust disease detection and grading | |
| Chowdhury et al. | Plant Leaf Disease Detection and Classification Using Deep Learning: A Review and A Proposed System on Bangladesh's Perspective | |
| Ali et al. | LodgeNet: an automated framework for precise detection and classification of wheat lodging severity levels in precision farming | |
| Malounas et al. | Testing the suitability of automated machine learning, hyperspectral imaging and CIELAB color space for proximal in situ fertilization level classification | |
| Menezes et al. | Pseudo-label semi-supervised learning for soybean monitoring | |
| CN119206489A (en) | An early identification and monitoring method for field crops based on three-stage ensemble learning | |
| Guzel et al. | Deep learning for image-based detection of weeds from emergence to maturity in wheat fields | |
| Khan et al. | A Systematic Literature Review of Machine Learning and Deep Learning Approaches for Spectral Image Classification in Agricultural Applications Using Aerial Photography. | |
| Adelaja et al. | Leveraging Deep Learning for Real-Time Coffee Leaf Disease Identification | |
| Valente et al. | Fast classification of large germinated fields via high-resolution UAV imagery | |
| Van Haeften et al. | Unmanned aerial vehicle phenotyping of agronomic and physiological traits in mungbean | |
| Nadiger et al. | Intelligent techniques for nutrition level prediction in guava plants | |
| Atianashie | Disease Detection in CCN-51 Cocoa Fruits through Convolutional Neural Networks: A Novel Approach for the Ghana Cocoa Board |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23751354 Country of ref document: EP Kind code of ref document: A1 |
|
| REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112024027616 Country of ref document: BR |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| REG | Reference to national code |
Ref country code: BR Ref legal event code: B01E Ref document number: 112024027616 Country of ref document: BR Free format text: APRESENTE NOVAS FOLHAS DO (RELATORIO DESCRITIVO) ADAPTADAS AO ART. 26. I DA PORTARIA/INPI/NO 14/2024, UMA VEZ QUE O TITULO ENVIADO NA PETICAO NO 870240111320 DE 30/12/2024 ENCONTRA-SE FORA DA NORMA NO QUE SE REFERE A FORMATACAO, MAIS ESPECIFICAMENTE NAO PODEM CONSTAR EXPRESSOES COMO "RELATORIO DESCRITIVO" OU OUTRAS JUNTO AO TITULO. DA MESMA FORMA APRESENTE NOVAS VIAS DO (RESUMO) ADAPTADAS AO ART. 18 DA PORTARIA/INPI/NO 14/2024. A EXIGENCIA DEVE SER RESPONDIDA EM ATE 60 (SESSENTA) DIAS DE SUA PUBLICACAO E DEVE SER REALIZADA POR MEIO DA PETICAO GRU CODIGO DE SERVICO 207. |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 23751354 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 112024027616 Country of ref document: BR Kind code of ref document: A2 Effective date: 20241230 |