[go: up one dir, main page]

US12389900B1 - Automatic animal detection and deterrent system - Google Patents

Automatic animal detection and deterrent system

Info

Publication number
US12389900B1
US12389900B1 US18/369,232 US202318369232A US12389900B1 US 12389900 B1 US12389900 B1 US 12389900B1 US 202318369232 A US202318369232 A US 202318369232A US 12389900 B1 US12389900 B1 US 12389900B1
Authority
US
United States
Prior art keywords
target
animal
camera
target animal
deterrent
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US18/369,232
Inventor
Charles Hartman King
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Western Robotics LLC
Original Assignee
Western Robotics LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Western Robotics LLC filed Critical Western Robotics LLC
Priority to US18/369,232 priority Critical patent/US12389900B1/en
Assigned to WESTERN ROBOTICS LLC reassignment WESTERN ROBOTICS LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Animal Deterrent Solutions, LLC
Application granted granted Critical
Publication of US12389900B1 publication Critical patent/US12389900B1/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M29/00Scaring or repelling devices, e.g. bird-scaring apparatus
    • A01M29/06Scaring or repelling devices, e.g. bird-scaring apparatus using visual means, e.g. scarecrows, moving elements, specific shapes, patterns or the like
    • A01M29/10Scaring or repelling devices, e.g. bird-scaring apparatus using visual means, e.g. scarecrows, moving elements, specific shapes, patterns or the like using light sources, e.g. lasers or flashing lights
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M29/00Scaring or repelling devices, e.g. bird-scaring apparatus
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M29/00Scaring or repelling devices, e.g. bird-scaring apparatus
    • A01M29/16Scaring or repelling devices, e.g. bird-scaring apparatus using sound waves
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M29/00Scaring or repelling devices, e.g. bird-scaring apparatus
    • A01M29/16Scaring or repelling devices, e.g. bird-scaring apparatus using sound waves
    • A01M29/18Scaring or repelling devices, e.g. bird-scaring apparatus using sound waves using ultrasonic signals
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01MCATCHING, TRAPPING OR SCARING OF ANIMALS; APPARATUS FOR THE DESTRUCTION OF NOXIOUS ANIMALS OR NOXIOUS PLANTS
    • A01M31/00Hunting appliances
    • A01M31/002Detecting animals in a given area
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands

Definitions

  • Pets such as dogs and cats, etc. have been known to cause harm to furnishings, plants and other objects by scratching, lying upon, and in general having access to such objects.
  • Predators such as coyotes are known to attack and kill pets such as dogs.
  • Marine animals such as birds and sea lions are known to perch or lie upon ocean craft in harbors and to soil and foul the surfaces of these craft.
  • devices are known to detect the presence of these animals and deter them by producing a stimulus which startles or scares the animal. These deterring stimuli can include the spraying the animal with water, emitting sounds, or flashing lights, or combinations thereof.
  • U.S. Pat. Nos. 4,658,386; 7,278,375; 7,462,364; 9,044,770; 9,204,622; and 9,226,493 disclose various methods and systems to detect and deter animals from entering a specified area. All of these methods and systems use infra-red motion detection systems to detect the presence of the animals. Motion detection has not been reliable and can lead to false positives and negatives. Motion detectors cannot distinguish between humans and non-human animals.
  • Pulse width modulated waves are generated for rotating a steering engine by using a discrete proportional integral derivative algorithm according to the information.
  • the pulse width modulated wave signals are transmitted to the steering engine so that the steering engine can align the flying birds to the center of the camera.
  • An instruction is then sent to a laser generator to emit lasers to frighten the birds away.
  • This invention provides a method of detecting and deterring a target animal from a target area.
  • the target area is placed within the field of vision of a camera which is connected to a computer processing system.
  • An animal identification computer program is run using convolution neural networks and deep learning computer programs with camera images from the camera to detect a target animal in the target area. It is verified whether a target animal is in the field of view of the camera. If the target animal is in the field of view of the camera the target animal is recorded with the camera.
  • One or more deterrents are aimed at the target animal and a target location is set. The deterrent is deployed to cause the target animal to leave the target area.
  • Another advantage is the ability to deploy a deterrent against a target animal within 0.25 to 2 seconds from the instant of detection so that little or no time is available to the target animal to damage the target area.
  • FIG. 4 is a flow chart of the steps of the method of the present invention for automatic detection and deterring of animals.
  • FIG. 1 shows a perspective view of the exterior of the automatic animal detection and deterrent system 10 of the present invention, which is contained in housing 11 having a front end 12 with a clear viewing window chamber 13 .
  • a video camera lens 14 is shown inside the clear viewing window chamber 13 .
  • a photovoltaic sensor (PV) 15 may also be contained within the clear viewing window chamber 13 to indicate day time or night time.
  • the electrical components of the housing 11 may be connected to a solar panel 16 and to a plurality of animal deterrent devices 17 , 18 , and 19 .
  • FIG. 2 is a diagram of the electrical components inside the housing 11 .
  • the lens 14 and the PV sensor 15 in the window chamber 13 are connected to the video camera 20 inside the housing 11 .
  • the camera 20 is attached to a central processing unit (CPU) 21 having a memory chip 22 which can collect data for datasets and store recordings from triggers.
  • An external computation virtual processing unit (VPU) 23 is connected to the central processing unit 21 , preferably by a USB connection.
  • the VPU is a hardware accelerator for deep neural network inferences to improve edge detection.
  • the CPU 21 is connected to a plurality of relays 24 , 25 , and 26 to transfer high voltage to the corresponding plurality of animal deterrents 17 , 18 , and 19 .
  • the video camera 20 monitors a specified area from which a user desires to deter unwanted animals. Data from the video camera 20 is transmitted to the central processing unit 21 which runs an animal identification program and controls all deterrent functions.
  • the CPU 21 can identify an unwanted animal of interest in the specified area, target the unwanted animal in the specified area, and deploy a deterrent to the unwanted animal which will cause the unwanted animal to leave the specified area.
  • the memory chip 22 contains identification and execution programs and stores recorded images of unwanted animals that are deterred.
  • the external central processing unit 23 assists in running computations and increases the performance of the CPU 21 .
  • the relays 24 , 25 , and 26 transfer necessary voltage to the animal deterrents 17 , 18 , and 19 , respectively, for example.
  • the animal deterrents can be sound devices, such as horns, sirens, and ultrasonic sound; light devices, such as flashing lights, strobe lights and laser lights; water spray devices, such as sprinklers, hoses, and water cannons; and compressed air used with projectiles, such as sand.
  • sound devices such as horns, sirens, and ultrasonic sound
  • light devices such as flashing lights, strobe lights and laser lights
  • water spray devices such as sprinklers, hoses, and water cannons
  • compressed air used with projectiles such as sand.
  • the voltage regulator automatically adjusts voltage from the battery 29 to power the CPU 21 and the VPU 23 .
  • the terminal block 29 distributes power from the battery 29 to the various electrical components in the system.
  • the charge controller 31 controls and balances charging levels to protect the battery 29 and other electrical components from electrically overloading or underloading.
  • the present invention uses convolution neural networks and deep learning technology known in the art as a means of identifying animal targets for deterring.
  • the technology allows identification of a specific species.
  • a large collection of images of targeted animals is obtained in a given environment.
  • a large collection of images of a given environment with no targeted animals therein is also obtained.
  • machine learning technology a set of characteristics are discovered that allow identification of a target animal. The identification is practically in real time with few false positives and is specific to an animal species.
  • FIG. 3 is a flow chart of a method of training the automatic animal detection and deterrent system to detect target animals.
  • the training process can begin with gathering target animal image data, preferably in an area where a camera will be deployed a (STEP 1).
  • An image data set is built for the target animal in the specific area (STEP 2).
  • image data is gathered at different times of the day using different image recording devices, and in different locations where a camera may be deployed.
  • Images in the data sets are annotated or labeled by methods known in the art so that the data sets may be entered into a standard learning algorithm (STEP 3).
  • the image size may be changed to model input parameter size. Distortion can be made to the data set to add more data for training (STEP 4).
  • the images may be divided into different sets for training, validation, and/or testing.
  • the algorithm is then trained on a training data set (STEP 5). Any known suitable learning algorithm may be used as desired.
  • the training results are evaluated on a validation data set (STEP 6). If validation results are acceptable the animal detection and deterrent system is ready to use (STEP 7). If the validation results are unacceptable the learning algorithm is retrained (STEP 8). Any new image data can be added to an existing database for any animal target.
  • FIG. 4 is a flow chart of the steps of the method of the present invention for automatic detection and deterring of animals.
  • the unit 10 is placed where it can monitor a target area wherein the target area is within the field of vision of the video camera (STEP 9).
  • the unit 10 is turned on (STEP 10) and the animal identification software program is running (STEP 11). Images received by the video camera are transmitted to the CPU which is programmed with a scanning identification algorithm that scans the real time video images for the pre-programmed target animal of interest. The scanning continues until the target animal is in the field of view and the target animal image is identified by the program (STEP 12).
  • the video images can be recorded if desired (STEP 14) while the deterrents are prepared for activation (STEP 15) and the target location is identified (STEP 16). One or more deterrents are then activated (STEP 17) and video recording is stopped (STEP 18) and saved to memory.
  • the program scans the video images to verify the presence or absence of the target animal (STEP 12). If so, STEPS 14-18 are repeated. If not, the animal identification program is continued (STEP 13) to monitor the presence or absence of the target animal.
  • the time between detection and deterrence is sub seconds or only a few seconds, for example, 0.25 to 2 seconds, which is sufficient to deter the target animal so that little or no damaged is produced by the target animal.
  • the system can be programmed and monitored with any suitable computer or related device, including, for example, cell phones.
  • the system can be pre-programmed (trained) to identify and deter as many animal types as desired. Specific individual humans can be identified as known or unknown. A plurality of different deterrents can be deployed at the same time when a target animal is detected.
  • Any type of camera can be used, including an IP camera. Multiple cameras can be used on one CPU. Connections to solar panels and deterrent devices may be wireless. Every trigger video as it happens, and/or daily summary of activity can be transmitted wirelessly via a cell network.

Landscapes

  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Pest Control & Pesticides (AREA)
  • Health & Medical Sciences (AREA)
  • Environmental Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Zoology (AREA)
  • Wood Science & Technology (AREA)
  • Insects & Arthropods (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Birds (AREA)
  • General Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Optics & Photonics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

This disclosure provides a method of detecting and deterring a target animal from a target area. A target area is positioned within the field of vision of a video camera connected to a computer processing system. An animal identification computer program using convolution neural networks and deep learning computer programs and camera images rapidly detects a target animal. The animal identification computer program is trained to identify target animals accurately using a learning algorithm and related machine learning technology. The time to deploy a deterrent against a target animal from the instant of detection is 2 seconds or less so that little or no time is available to the target animal to damage the target area.

Description

CROSS REFERENCE TO RELATED APPLICATIONS
This patent application is a continuation of, and claims the benefit of, U.S. patent application Ser. No. 17/748,118, currently pending, which is a continuation of, and claims the benefit of U.S. Pat. No. 11,369,106 and which are hereby incorporated herein by reference in their entirety.
FIELD OF THE INVENTION
This invention relates to methods and systems for deterring animals from entering specific areas and, more particularly, to methods and systems that use video object detection using machine learning on a neural network and type identification training to quickly, accurately, and automatically detect and deter an intruding animal in a given location.
BACKGROUND OF THE INVENTION
It has been desirable to exclude animals from a particular area. Pets, such as dogs and cats, etc. have been known to cause harm to furnishings, plants and other objects by scratching, lying upon, and in general having access to such objects.
Predators such as coyotes are known to attack and kill pets such as dogs. Marine animals such as birds and sea lions are known to perch or lie upon ocean craft in harbors and to soil and foul the surfaces of these craft. To combat such unwanted behavior, devices are known to detect the presence of these animals and deter them by producing a stimulus which startles or scares the animal. These deterring stimuli can include the spraying the animal with water, emitting sounds, or flashing lights, or combinations thereof. U.S. Pat. Nos. 4,658,386; 7,278,375; 7,462,364; 9,044,770; 9,204,622; and 9,226,493 disclose various methods and systems to detect and deter animals from entering a specified area. All of these methods and systems use infra-red motion detection systems to detect the presence of the animals. Motion detection has not been reliable and can lead to false positives and negatives. Motion detectors cannot distinguish between humans and non-human animals.
China Patent Application Publication No. CN202476328 discloses an airport laser bird repelling system based on image recognition (classification). A motion detection module firstly uses the images of a front frame and a rear frame of a video stream as differential, so as to find an object in motion. A bird identification module is the used for identifying the image morphology of all the moving objects. When the characteristic of the image changes regularly, namely the movement of the flying bird wings, the object is considered as “birds”. A movement tracking module tracks the “birds” by using a tracking algorithm. A position parameter generating module is used for converting the cartesian coordinate system position information on the image into polar coordinate system information, namely the distance from the center of the image and the rotation angle around the initial axis. Pulse width modulated waves are generated for rotating a steering engine by using a discrete proportional integral derivative algorithm according to the information. The pulse width modulated wave signals are transmitted to the steering engine so that the steering engine can align the flying birds to the center of the camera. An instruction is then sent to a laser generator to emit lasers to frighten the birds away. Although this use of video image recognition may improve upon detection of only motion, object recognition via morphology is relatively slow, complex, inaccurate, and imprecise.
SUMMARY OF THE INVENTION
This invention provides a method of detecting and deterring a target animal from a target area. The target area is placed within the field of vision of a camera which is connected to a computer processing system. An animal identification computer program is run using convolution neural networks and deep learning computer programs with camera images from the camera to detect a target animal in the target area. It is verified whether a target animal is in the field of view of the camera. If the target animal is in the field of view of the camera the target animal is recorded with the camera. One or more deterrents are aimed at the target animal and a target location is set. The deterrent is deployed to cause the target animal to leave the target area.
The recording is stopped after the deterrent is deployed, and the recording is saved to a file in the computer. If the target animal has not left the target area after deploying the deterrent the process is repeated. Verification whether a target animal is in the field of vision of the camera is performed continuously until a target animal is in the field of view of the camera. The animal identification computer program is, preferably, trained to identify target animals using a learning algorithm which involves training the learning algorithm with training data sets. The training of the learning algorithm with training data sets is validated with validation data sets. The training data sets and the validation data sets are created by gathering target animal image data in a target area, building image data sets for target animals and for target areas from the image data, and annotating or labeling images in the data sets so that the data sets may be entered into the learning algorithm.
An advantage of the method of the present invention is detecting target animals of specific species from a target area using convolution neural networks and deep learning technology to identify animal targets rapidly.
Another advantage is video camera identification of a target animal regardless of the background of the video.
Another advantage is the use of machine learning technology to provide highly accurate identification of target animals with few or no false positives.
Another advantage is the ability to deploy a deterrent against a target animal within 0.25 to 2 seconds from the instant of detection so that little or no time is available to the target animal to damage the target area.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 shows a perspective view of the exterior of an automatic animal detection and deterrent system of the present invention.
FIG. 2 is a diagram of the electro-mechanical components of the automatic animal detection and deterrent system.
FIG. 3 is a flow chart of the method of training the automatic animal detection and deterrent system to detect objects using object detection through machine learning on a neural network.
FIG. 4 is a flow chart of the steps of the method of the present invention for automatic detection and deterring of animals.
DESCRIPTION OF THE INVENTION
While the following description details the preferred embodiments of the present invention, it is to be understood that the invention is not limited in its application to the details of arrangement of the parts or steps of the methods illustrated in the accompanying figures, since the invention is capable of other embodiments and of being practiced in various ways.
FIG. 1 shows a perspective view of the exterior of the automatic animal detection and deterrent system 10 of the present invention, which is contained in housing 11 having a front end 12 with a clear viewing window chamber 13. A video camera lens 14 is shown inside the clear viewing window chamber 13. A photovoltaic sensor (PV) 15 may also be contained within the clear viewing window chamber 13 to indicate day time or night time. The electrical components of the housing 11 may be connected to a solar panel 16 and to a plurality of animal deterrent devices 17, 18, and 19.
FIG. 2 is a diagram of the electrical components inside the housing 11. The lens 14 and the PV sensor 15 in the window chamber 13 are connected to the video camera 20 inside the housing 11. The camera 20 is attached to a central processing unit (CPU) 21 having a memory chip 22 which can collect data for datasets and store recordings from triggers. An external computation virtual processing unit (VPU) 23 is connected to the central processing unit 21, preferably by a USB connection. The VPU is a hardware accelerator for deep neural network inferences to improve edge detection. The CPU 21 is connected to a plurality of relays 24, 25, and 26 to transfer high voltage to the corresponding plurality of animal deterrents 17, 18, and 19. The CPU 21 receives power through a voltage regulator 27 which receives power through a terminal block 28 from a battery 29. A charge controller 30 is connected to the battery 29 and to the terminal block 28 through an on/off switch 31. The charge controller 30 may receive power from the solar panel 16 or other external sources of electrical power.
The video camera 20 monitors a specified area from which a user desires to deter unwanted animals. Data from the video camera 20 is transmitted to the central processing unit 21 which runs an animal identification program and controls all deterrent functions. The CPU 21 can identify an unwanted animal of interest in the specified area, target the unwanted animal in the specified area, and deploy a deterrent to the unwanted animal which will cause the unwanted animal to leave the specified area. The memory chip 22 contains identification and execution programs and stores recorded images of unwanted animals that are deterred. The external central processing unit 23 assists in running computations and increases the performance of the CPU 21. The relays 24, 25, and 26 transfer necessary voltage to the animal deterrents 17, 18, and 19, respectively, for example. The animal deterrents, without limitation, can be sound devices, such as horns, sirens, and ultrasonic sound; light devices, such as flashing lights, strobe lights and laser lights; water spray devices, such as sprinklers, hoses, and water cannons; and compressed air used with projectiles, such as sand.
The voltage regulator automatically adjusts voltage from the battery 29 to power the CPU 21 and the VPU 23. The terminal block 29 distributes power from the battery 29 to the various electrical components in the system. The charge controller 31 controls and balances charging levels to protect the battery 29 and other electrical components from electrically overloading or underloading.
The present invention uses convolution neural networks and deep learning technology known in the art as a means of identifying animal targets for deterring. The technology allows identification of a specific species. There are several commercially available pre-trained object detection models, but they are insufficient for detecting an animal. It is essential that the target animal can be identified in the video regardless of the background of the video so that identification can occur quickly before the target animal can do any damage. A large collection of images of targeted animals is obtained in a given environment. A large collection of images of a given environment with no targeted animals therein is also obtained. Using machine learning technology, a set of characteristics are discovered that allow identification of a target animal. The identification is practically in real time with few false positives and is specific to an animal species.
FIG. 3 is a flow chart of a method of training the automatic animal detection and deterrent system to detect target animals. The training process can begin with gathering target animal image data, preferably in an area where a camera will be deployed a (STEP 1). An image data set is built for the target animal in the specific area (STEP 2). Preferably, image data is gathered at different times of the day using different image recording devices, and in different locations where a camera may be deployed. Images in the data sets are annotated or labeled by methods known in the art so that the data sets may be entered into a standard learning algorithm (STEP 3). The image size may be changed to model input parameter size. Distortion can be made to the data set to add more data for training (STEP 4). The images may be divided into different sets for training, validation, and/or testing. The algorithm is then trained on a training data set (STEP 5). Any known suitable learning algorithm may be used as desired. The training results are evaluated on a validation data set (STEP 6). If validation results are acceptable the animal detection and deterrent system is ready to use (STEP 7). If the validation results are unacceptable the learning algorithm is retrained (STEP 8). Any new image data can be added to an existing database for any animal target.
FIG. 4 is a flow chart of the steps of the method of the present invention for automatic detection and deterring of animals. Once the automatic animal detection and deterrent system has been preprogrammed with the training sequence for one or more target animals it can be placed in a running mode. The unit 10 is placed where it can monitor a target area wherein the target area is within the field of vision of the video camera (STEP 9). The unit 10 is turned on (STEP 10) and the animal identification software program is running (STEP 11). Images received by the video camera are transmitted to the CPU which is programmed with a scanning identification algorithm that scans the real time video images for the pre-programmed target animal of interest. The scanning continues until the target animal is in the field of view and the target animal image is identified by the program (STEP 12). The video images can be recorded if desired (STEP 14) while the deterrents are prepared for activation (STEP 15) and the target location is identified (STEP 16). One or more deterrents are then activated (STEP 17) and video recording is stopped (STEP 18) and saved to memory. The program scans the video images to verify the presence or absence of the target animal (STEP 12). If so, STEPS 14-18 are repeated. If not, the animal identification program is continued (STEP 13) to monitor the presence or absence of the target animal. The time between detection and deterrence is sub seconds or only a few seconds, for example, 0.25 to 2 seconds, which is sufficient to deter the target animal so that little or no damaged is produced by the target animal.
The foregoing description has been limited to specific embodiments of this invention. It will be apparent, however, that variations and modifications may be made by those skilled in the art to the disclosed embodiments of the invention, with the attainment of some or all of its advantages and without departing from the spirit and scope of the present invention. For example, the system can be programmed and monitored with any suitable computer or related device, including, for example, cell phones. The system can be pre-programmed (trained) to identify and deter as many animal types as desired. Specific individual humans can be identified as known or unknown. A plurality of different deterrents can be deployed at the same time when a target animal is detected. Any type of camera can be used, including an IP camera. Multiple cameras can be used on one CPU. Connections to solar panels and deterrent devices may be wireless. Every trigger video as it happens, and/or daily summary of activity can be transmitted wirelessly via a cell network.
It will be understood that various changes in the details, materials, and arrangements of the parts which have been described and illustrated above in order to explain the nature of this invention may be made by those skilled in the art without departing from the principle and scope of the invention as recited in the following claims.

Claims (6)

The invention claimed is:
1. A method of detecting and deterring a target animal from a target area, comprising:
1) placing the target area within the field of vision of a camera, wherein the camera is connected to a computer processing system;
2) running an animal identification computer program with camera images from the camera to detect a target animal;
3) training the animal identification computer program to identify target animals using a learning algorithm;
4) deploying a deterrent to cause the target animal to leave the target area within 2 seconds from the instant of detection; and
5) repeating step 4) if the target animal has not left the target area after deploying the deterrent,
wherein convolution neural networks and deep learning computer programs with the camera images are used to detect the target animal,
wherein the target animal in the field of view of the camera is verified,
wherein one or more deterrents are armed, and a target location is set, and
wherein the species of the target animal is identified.
2. The method of claim 1, further comprising recording the target animal with the camera in the field of view of the camera.
3. The method of claim 1, wherein the deterrent is a sound device, a light device, a spray device, projectiles, or combinations thereof.
4. A method of detecting and deterring a target animal from a target area, comprising:
1) placing the target area within the field of vision of a camera, wherein the camera is connected to a computer processing system;
2) running an animal identification computer program with camera images from the camera to detect a target animal;
3) training the animal identification computer program to identify target animals using a learning algorithm;
4) verifying the target animal in the field of view of the camera;
5) recording the target animal with the camera if the target animal is in the field of view of the camera;
6) arming one or more deterrents and setting a target location;
7) deploying the deterrent against the target animal within 2 seconds from the instant of detection; and
8) repeating step 7) if the target animal has not left the target area after deploying the deterrent,
wherein convolution neural networks and deep learning computer programs with the camera images are used to detect the target animal, and
wherein the species of the target animal is identified.
5. The method of claim 4, wherein the deterrent is a sound device, a light device, a spray device, projectiles, or combinations thereof.
6. A method of detecting and deterring a target animal from a target area, comprising:
1) placing the target area within the field of vision of a camera, wherein the camera is connected to a computer processing system;
2) running an animal identification computer program with camera images from the camera to detect a target animal;
3) training the animal identification computer program to identify target animals using a learning algorithm;
4) verifying the target animal in the field of view of the camera;
5) recording the target animal with the camera if the target animal is in the field of view of the camera;
6) identifying the species of the target animal;
7) arming one or more deterrents and setting a target location, wherein the deterrent is a sound device, a light device, a spray device, projectiles, or combinations thereof;
8) deploying the deterrent against the target animal within 2 seconds from the instant of detection; and
9) repeating step 8) if the target animal has not left the target area after deploying the deterrent,
wherein convolution neural networks and deep learning computer programs with the camera images are used to detect the target animal.
US18/369,232 2019-04-12 2023-09-18 Automatic animal detection and deterrent system Active 2039-09-22 US12389900B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/369,232 US12389900B1 (en) 2019-04-12 2023-09-18 Automatic animal detection and deterrent system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US16/382,498 US11369106B2 (en) 2019-04-12 2019-04-12 Automatic animal detection and deterrent system
US17/748,118 US11793190B1 (en) 2019-04-12 2022-05-19 Automatic animal detection and deterrent system
US18/369,232 US12389900B1 (en) 2019-04-12 2023-09-18 Automatic animal detection and deterrent system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US17/748,118 Continuation US11793190B1 (en) 2019-04-12 2022-05-19 Automatic animal detection and deterrent system

Publications (1)

Publication Number Publication Date
US12389900B1 true US12389900B1 (en) 2025-08-19

Family

ID=72749447

Family Applications (3)

Application Number Title Priority Date Filing Date
US16/382,498 Active 2041-04-28 US11369106B2 (en) 2019-04-12 2019-04-12 Automatic animal detection and deterrent system
US17/748,118 Active 2039-04-18 US11793190B1 (en) 2019-04-12 2022-05-19 Automatic animal detection and deterrent system
US18/369,232 Active 2039-09-22 US12389900B1 (en) 2019-04-12 2023-09-18 Automatic animal detection and deterrent system

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US16/382,498 Active 2041-04-28 US11369106B2 (en) 2019-04-12 2019-04-12 Automatic animal detection and deterrent system
US17/748,118 Active 2039-04-18 US11793190B1 (en) 2019-04-12 2022-05-19 Automatic animal detection and deterrent system

Country Status (1)

Country Link
US (3) US11369106B2 (en)

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB201519845D0 (en) * 2015-11-10 2015-12-23 Rat Tec Solutions Ltd Animal sorting device
BR112021002477A2 (en) * 2018-09-21 2021-07-27 Bayer Aktiengesellschaft sensor assisted image
US11109586B2 (en) * 2019-11-13 2021-09-07 Bird Control Group, Bv System and methods for automated wildlife detection, monitoring and control
US20210315186A1 (en) * 2020-04-14 2021-10-14 The United States Of America, As Represented By Secretary Of Agriculture Intelligent dual sensory species-specific recognition trigger system
US11991961B2 (en) * 2020-08-18 2024-05-28 The Phoebus Fund, LLC Solar power generation and agricultural material dispersal system
CN116018064A (en) * 2020-09-03 2023-04-25 昕诺飞控股有限公司 Method and system for expelling predators of fish
EP4208020A4 (en) * 2020-09-03 2024-11-27 Proa Holdings Pty Ltd SYSTEM AND METHOD FOR WILDLIFE DETECTION AND DETERRENCE FOR ENVIRONMENTAL INSTRUMENTATION
JP7166655B2 (en) * 2020-10-19 2022-11-08 新生電子株式会社 bird repellent device
CN112257673B (en) * 2020-11-17 2025-02-28 携程计算机技术(上海)有限公司 Animal identification method, system, device and storage medium based on tourism images
CN113287597B (en) * 2021-05-20 2023-04-18 河南天通电力有限公司 Transmission line initiative bird repellent device based on video is studied and judged
CN113642111A (en) * 2021-08-20 2021-11-12 Oppo广东移动通信有限公司 Safety protection method and device, medium, electronic equipment and vehicle
CN113875738B (en) * 2021-09-06 2023-05-05 滨州凯飞电力器材有限公司 Array type laser bird repellent device and control system
GB2614708A (en) 2022-01-10 2023-07-19 Sony Semiconductor Solutions Corp An animal detecting device, system, method and computer program
EP4324328A1 (en) 2022-08-19 2024-02-21 CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement Autonomous anti-collision system
CN116267880A (en) * 2023-03-13 2023-06-23 深圳特立克科技有限公司 A kind of AI intelligent sound wave pulse resonance mouse repeller and mouse repelling method
CN117441701B (en) * 2023-10-25 2025-11-04 常州大学 A UAV-based method and system for bird control in agriculture based on topological sorting reward mechanism
KR102805949B1 (en) * 2023-12-15 2025-05-14 주식회사 다운 Device for eradicating harmful birds using AI technology
CN117690164B (en) * 2024-01-30 2024-04-30 成都欣纳科技有限公司 Airport bird identification and driving method and system based on edge calculation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160133131A1 (en) * 2014-11-12 2016-05-12 GM Global Technology Operations LLC Use of participative sensing systems to enable enhanced road friction estimation
US20190166823A1 (en) * 2017-12-04 2019-06-06 Caldera Services LLC Selective Action Animal Trap
US20190246623A1 (en) * 2016-07-08 2019-08-15 Commonwealth Scientific And Industrial Research Organisation Pest deterrent system
US20210324832A1 (en) * 2014-08-21 2021-10-21 Identiflight International, Llc Imaging Array for Bird or Bat Detection and Identification

Family Cites Families (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN201726780U (en) 2010-06-11 2011-02-02 中电国科(北京)科技有限公司 Intelligent bird-expelling system for transmission line
US20140261151A1 (en) * 2011-09-23 2014-09-18 Lite Enterprise, Inc. Method and system for provoking an avoidance behavioral response in animals
US8988230B2 (en) 2011-10-25 2015-03-24 Accipiter Radar Technologies Inc. Device and method for smart, non-habituating, automatic bird deterrent system
CN202354274U (en) 2011-10-26 2012-08-01 北京友泰顺城科技发展有限公司 Laser bird scaring device
CN202476328U (en) 2012-02-21 2012-10-10 中国民航大学 Image recognition-based airport laser bird-repelling system
CN202890329U (en) 2012-09-20 2013-04-24 西南科技大学 Automatic aiming laser bird repellent device based on pattern recognition
CN102860300B (en) 2012-09-20 2014-04-02 西南科技大学 A pattern recognition-based automatic aiming laser bird repelling device and bird repelling method
US9226493B2 (en) 2012-11-02 2016-01-05 Mesa Digital, Llc Varmint and intruder deterrent system
US9775337B2 (en) * 2012-11-27 2017-10-03 Elwha Llc Methods and systems for directing birds away from equipment
US9762865B2 (en) 2013-03-15 2017-09-12 James Carey Video identification and analytical recognition system
US9295225B2 (en) 2013-03-15 2016-03-29 Harold G Monk Species specific feeder
US9424461B1 (en) 2013-06-27 2016-08-23 Amazon Technologies, Inc. Object recognition for three-dimensional bodies
ES2765257T3 (en) * 2013-12-19 2020-06-08 Bird Control Group B V Bird deterrence system
US10280402B2 (en) 2014-08-06 2019-05-07 College Of Medicine Pochon Cha University Industry-Academic Cooperation Foundation Immune-compatible cells created by nuclease-mediated editing of genes encoding HLA
US9563825B2 (en) 2014-11-20 2017-02-07 Adobe Systems Incorporated Convolutional neural network using a binarized convolution layer
US9652838B1 (en) 2014-12-23 2017-05-16 A9.Com, Inc. Object retrieval
WO2017020060A1 (en) * 2015-08-05 2017-02-09 Ecological Horizons Pty Ltd Automated device for delivering a pharmaceutical to an animal
KR102147361B1 (en) 2015-09-18 2020-08-24 삼성전자주식회사 Method and apparatus of object recognition, Method and apparatus of learning for object recognition
US10169684B1 (en) 2015-10-01 2019-01-01 Intellivision Technologies Corp. Methods and systems for recognizing objects based on one or more stored training images
US9959468B2 (en) 2015-11-06 2018-05-01 The Boeing Company Systems and methods for object tracking and classification
CN106070175A (en) 2016-06-29 2016-11-09 国网山东省电力公司济南市历城区供电公司 A kind of modified model bird-scaring unit
US10290196B2 (en) 2016-08-15 2019-05-14 Nec Corporation Smuggling detection system
GB2555836A (en) * 2016-11-11 2018-05-16 Bioseco Sp Z O O Systems and methods for detecting flying animals
US10694737B2 (en) * 2017-05-12 2020-06-30 The Boeing Company Keeping animals from protected areas using eye-safe lasers
US20190152595A1 (en) * 2017-11-17 2019-05-23 Bigfoot Technologies Inc. Apparatus for Sustained Surveillance and Deterrence with Unmanned Aerial Vehicles (UAV)

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210324832A1 (en) * 2014-08-21 2021-10-21 Identiflight International, Llc Imaging Array for Bird or Bat Detection and Identification
US20160133131A1 (en) * 2014-11-12 2016-05-12 GM Global Technology Operations LLC Use of participative sensing systems to enable enhanced road friction estimation
US20190246623A1 (en) * 2016-07-08 2019-08-15 Commonwealth Scientific And Industrial Research Organisation Pest deterrent system
US20190166823A1 (en) * 2017-12-04 2019-06-06 Caldera Services LLC Selective Action Animal Trap

Also Published As

Publication number Publication date
US11369106B2 (en) 2022-06-28
US11793190B1 (en) 2023-10-24
US20200323193A1 (en) 2020-10-15

Similar Documents

Publication Publication Date Title
US12389900B1 (en) Automatic animal detection and deterrent system
US11617363B2 (en) Robotic agriculture protection system
Fenton Echolocation: implications for ecology and evolution of bats
KR20180001429A (en) Vermin Prevention System based on IoT Technology with Sound Detection
US20210084888A1 (en) Intelligent Control System for Deterring Problem Animals
CN120304395A (en) A method for driving away birds from power distribution tower and related equipment
US20210315186A1 (en) Intelligent dual sensory species-specific recognition trigger system
AU2020103507A4 (en) System and method for crop protection from animals using smart sensors
KR102421330B1 (en) A device to repel tracked wildlife animal to prevent damage to crops
Rani et al. AI based scarecrow preventing from crop vandalization
CN115695732A (en) Method and system for monitoring state of high-voltage transmission line and storage medium
CN106172358A (en) Drive the method for birds, server and information collecting device
JP6183734B1 (en) Method and structure for preventing flight of harmful birds such as starlings and sparrows
US20070163516A1 (en) Intelligent scarecrow system for utilization in agricultural and industrial applications
CN119344292A (en) Intelligent bird repellent device and repellent strategy generation method for power transmission and distribution facilities
KR102249642B1 (en) A method of exterminating harmful animals, and apparatuses performing the same
Mahmood et al. Uav based smart bird control using convolutional neural networks
Nishanth et al. YoloV8-Powered Surveillance: Safeguarding Crops from Wildlife and Fire Hazards
Sannakashappanavar et al. Intelligent Surveillance and Protection System for Farmlands from Animals
JP7228852B6 (en) Bird and beast intimidation system
US20230329219A1 (en) System and Methods for Mitigation of Dangerous Wildlife
Akula et al. A comparative study and analysis of approaches towards agricultural supervision
JP2024024586A (en) Bird and beast damage control device
EP4324328A1 (en) Autonomous anti-collision system
Dadhich Increasing the accuracy of rodent detection and estimation of the population with sensor fusion

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE