Disclosure of Invention
Aiming at the defects in the prior art, the invention aims to provide a personnel overboard alarm system and a personnel overboard alarm method based on a convolutional neural network and image fusion.
The invention provides a personnel overboard alarm system based on a convolutional neural network and image fusion, which comprises:
an image acquisition module: the image acquisition module acquires data information of visible light images and infrared images;
an image storage module: the image storage module stores the visible light image and the infrared image acquired by the image acquisition module in real time and compresses the visible light image and the infrared image into a video stream;
an image registration module: the image registration module calibrates the spatial information of the visible light image and the infrared image;
an image fusion module: fusing the visible light image and the infrared image calibrated by the image registration module by utilizing a multi-scale transformation and fusion rule;
a convolutional neural network module: the convolutional neural network module performs target detection by using the fused visible light image and infrared image, and judges whether a dangerous case of people falling into water occurs;
the control alarm module: feeding back a judgment result of the convolutional neural network module to a control alarm module, starting an alarm when a dangerous case occurs, and storing water falling information;
an image display module: and the image display module displays the output result of the convolutional neural network module and the data of the image storage module.
Preferably, the image registration module comprises: the image registration module carries out smooth denoising processing on the visible light image and the infrared image, and calibrates the visible light image and the infrared image at the same moment according to the field angle range of the image resolution, so that the spatial areas of the images are the same.
Preferably, the convolutional neural network module includes: the method comprises the steps that a deep learning model trained by visible light image data and infrared image data is utilized, real-time image data are input into the deep learning model, whether a current image contains people falling into water or not is calculated according to network weights obtained by training, and if the current image contains people falling into water, marking is carried out;
the convolutional neural network model takes a residual error network as a main body, and a neural network layer comprises a convolutional layer and a pooling layer;
the deep learning model is a convolutional neural network model;
when the deep learning model is trained by using the visible light image data and the infrared image data, the visible light image data and the infrared image data comprise image data of people who fall into water and image data of people who do not fall into water, and the quantity of the image data of the people who fall into water is equivalent to that of the image data of the people who do not fall into water.
Preferably, the control alarm module comprises: the alarm device comprises an alarm device unit and an alarm information storage unit;
when the control alarm module obtains the dangerous case information, the alarm unit starts an alarm; the alarm information storage unit stores the drowning information;
the drowning information comprises a drowning time and a drowning position.
Preferably, the image display module includes: displaying different image data according to different output results of the convolutional neural network module, and when an emergency occurs, displaying data of an emergency area and marking people falling into water; when no dangerous case occurs, displaying real-time images of each monitoring area;
the image display module displays the data of the image storage module, and the video is read from the image storage module in a selected time period.
The invention provides a personnel overboard alarm method based on a convolutional neural network and image fusion, which comprises the following steps:
an image acquisition step: the image acquisition module acquires data information of visible light images and infrared images;
an image storage step: the image storage module stores the visible light image and the infrared image acquired in the image acquisition step in real time and compresses the visible light image and the infrared image into a video stream;
an image registration step: the image registration module calibrates the spatial information of the visible light image and the infrared image;
an image fusion step: fusing the visible light image and the infrared image calibrated in the image registration step by utilizing a multi-scale transformation and fusion rule;
a convolution neural network step: the convolutional neural network module performs target detection by using the fused visible light image and infrared image, and judges whether a dangerous case of people falling into water occurs;
controlling and alarming: feeding back a judgment result of the convolutional neural network module to a control alarm module, starting an alarm when a dangerous case occurs, and storing water falling information;
an image display module: and the image display module displays the output result of the convolutional neural network module and the data of the image storage module.
Preferably, the image registration step comprises: the image registration module carries out smooth denoising processing on the visible light image and the infrared image, and calibrates the visible light image and the infrared image at the same moment according to the field angle range of the image resolution, so that the spatial areas of the images are the same.
Preferably, the convolutional neural network step comprises: the method comprises the steps that a deep learning model trained by visible light image data and infrared image data is utilized, real-time image data are input into the deep learning model, whether a current image contains people falling into water or not is calculated according to network weights obtained by training, and if the current image contains people falling into water, marking is carried out;
the convolutional neural network model takes a residual error network as a main body, and a neural network layer comprises a convolutional layer and a pooling layer;
the deep learning model is a convolutional neural network model;
when the deep learning model is trained by using the visible light image data and the infrared image data, the visible light image data and the infrared image data comprise image data of people who fall into water and image data of people who do not fall into water, and the quantity of the image data of the people who fall into water is equivalent to that of the image data of the people who do not fall into water.
Preferably, the controlling and alarming step includes: the control alarm module comprises an alarm unit and an alarm information storage unit; when the control alarm module obtains the dangerous case information, the alarm unit starts an alarm; the alarm information storage unit stores the drowning information;
the drowning information comprises a drowning time and a drowning position.
Preferably, the image displaying step includes: displaying different image data according to different output results of the convolutional neural network module, and when an emergency occurs, displaying data of an emergency area and marking people falling into water; when no dangerous case occurs, displaying real-time images of each monitoring area;
the image display module displays the data of the image storage module, and the video is read from the image storage module in a selected time period.
Compared with the prior art, the invention has the following beneficial effects:
1. by utilizing the characteristics of high resolution of the visible light camera and all-weather monitoring of the infrared imager, the visible light camera has strong detail resolution capability, and pixels can reach millions. Designing a data fusion algorithm to obtain a fusion image of the two devices;
2. establishing navigation environment databases of two imaging devices, analyzing image characteristics of navigation environments, designing a neural network structure, establishing a target detection model, and realizing end-to-end training on acquired navigation pictures, wherein the obtained deep learning model can overcome the defects of the traditional target detection algorithm, has better flexibility and more accurate result;
3. when the dangerous case that people fall into water is detected, automatic alarm is realized, workers are reminded to intervene in time, and the falling water information is stored in a database so as to carry out detailed analysis and history tracing.
4. The method has the advantages that the higher drowning person identification rate is obtained in the complex navigation environment compared with the traditional target detection algorithm.
Detailed Description
The present invention will be described in detail with reference to specific examples. The following examples will assist those skilled in the art in further understanding the invention, but are not intended to limit the invention in any way. It should be noted that it would be obvious to those skilled in the art that various changes and modifications can be made without departing from the spirit of the invention. All falling within the scope of the present invention.
Aiming at the technical problems in the prior art, the method provided by the invention processes the visible light image and the infrared image data based on a deep learning method so as to realize personnel target identification, and simultaneously develops an image fusion algorithm by utilizing the complementary characteristics of two devices, enhances the image characteristic information and improves the identification accuracy.
The invention provides a personnel overboard alarm system based on a convolutional neural network and image fusion, which comprises: as shown in figure 1 of the drawings, in which,
an image acquisition module: the image acquisition module acquires data information of visible light images and infrared images;
the image acquisition module is divided into a visible light acquisition device and an infrared image acquisition device, and image data output by different image acquisition devices are subjected to interframe synchronization. The image storage module compresses the collected image data into a video data stream in an H.264 format and stores the video data stream in a disk array;
the image acquisition module is respectively connected with the image storage module and the image registration module, and the image storage module stores real-time image data as a video stream in an H.264 format.
An image storage module: the image storage module stores the visible light image and the infrared image acquired by the image acquisition module in real time and compresses the visible light image and the infrared image into a video stream;
an image registration module: the image registration module calibrates the spatial information of the visible light image and the infrared image;
specifically, the image registration module comprises: the image registration module carries out smooth denoising processing on the visible light image and the infrared image, and calibrates the visible light image and the infrared image at the same moment according to the field angle range of the image resolution, so that the spatial areas of the images are the same.
An image fusion module: the visible light image and the infrared image which are calibrated by the image registration module are fused by utilizing a multi-scale transformation and fusion rule to realize pixel-level fusion; in the embodiment, the visible light and infrared image fusion method is mainly applied to the collection of images under the condition of variable environment, so that the identification degree of the fused images is higher;
a convolutional neural network module: the convolutional neural network module performs target detection by using the fused visible light image and infrared image, and judges whether a dangerous case of people falling into water occurs;
specifically, the convolutional neural network module includes: the deep learning model trained by utilizing visible light image data and infrared image data is actually used for collecting pictures of a person drowning scene by deploying visible light and infrared photographic equipment, and on the data set, a neural network structure for target detection is designed to establish a two-classification deep learning model suitable for a ship navigation environment.
The convolutional neural network module realizes the feature extraction of the fused image, outputs a target detection result in a specific layer after multi-layer convolution and activation function, and feeds the result back to the control alarm module;
inputting real-time image data into a deep learning model, calculating whether a current image contains a person falling into water or not according to the network weight obtained by training, and if the current image contains the person falling into water, marking;
the convolutional neural network model takes a residual error network as a main body, and a neural network layer comprises a convolutional layer and a pooling layer;
the deep learning model is a convolutional neural network model;
when the deep learning model is trained by using the visible light image data and the infrared image data, the visible light image data and the infrared image data comprise image data of people who fall into water and image data of people who do not fall into water, and the quantity of the image data of the people who fall into water is equivalent to that of the image data of the people who do not fall into water.
In the present embodiment, the convolutional neural network is mainly composed of a residual network, and is composed of a plurality of network layers, including convolutional layers, pooling layers, and the like, and adopts a full convolutional layer structure, and the size of an input image is variable;
the control alarm module: the judgment result of the convolutional neural network module is fed back to the control alarm module, when a dangerous case occurs, the background alarm is controlled to be turned on or off, and the water falling information is stored so as to facilitate history tracing;
specifically, the control alarm module includes: the alarm device comprises an alarm device unit and an alarm information storage unit;
when the control alarm module obtains the dangerous case information, the alarm unit starts an alarm; reminding workers to intervene, storing the information falling into water by the alarm information storage unit, storing the information falling into water into the SQLITE database, calling real-time images of the image acquisition module by the workers in time for intervention, calling playback video segments of the video storage module, and reasonably deploying rescue actions after detailed judgment. The historical data can be conveniently traced;
the drowning information comprises a drowning time and a drowning position.
An image display module: and the image display module displays the output result of the convolutional neural network module and the historical data of the image storage module.
Specifically, the image display module includes: displaying different image data according to different output results of the convolutional neural network module, and when an emergency occurs, displaying data of an emergency area and marking people falling into water; when no dangerous case occurs, displaying real-time images of each monitoring area;
meanwhile, a history playback function is realized, the image display module displays data of the image storage module, and a worker can read videos from the image storage module at a selected time period.
The invention provides a personnel overboard alarm method based on a convolutional neural network and image fusion, which comprises the following steps: as shown in figure 1 of the drawings, in which,
an image acquisition step: the image acquisition module acquires data information of visible light images and infrared images;
the image acquisition step is divided into a visible light acquisition device and an infrared image acquisition device, and image data output by different image acquisition devices are subjected to interframe synchronization. The image storage module compresses the collected image data into a video data stream in an H.264 format and stores the video data stream in a disk array;
the image acquisition module is respectively connected with the image storage module and the image registration module, and the image storage module stores real-time image data as a video stream in an H.264 format.
An image storage step: the image storage module stores the visible light image and the infrared image acquired by the image acquisition module in real time and compresses the visible light image and the infrared image into a video stream;
an image registration step: the image registration module calibrates the spatial information of the visible light image and the infrared image;
specifically, the image registration step includes: the image registration module carries out smooth denoising processing on the visible light image and the infrared image, and calibrates the visible light image and the infrared image at the same moment according to the field angle range of the image resolution, so that the spatial areas of the images are the same.
An image fusion step: the visible light image and the infrared image which are calibrated by the image registration module are fused by utilizing a multi-scale transformation and fusion rule to realize pixel-level fusion; in the embodiment, the visible light and infrared image fusion method is mainly applied to the collection of images under the condition of variable environment, so that the identification degree of the fused images is higher;
a convolution neural network step: the convolutional neural network module performs target detection by using the fused visible light image and infrared image, and judges whether a dangerous case of people falling into water occurs;
specifically, the convolutional neural network step includes: the deep learning model trained by utilizing visible light image data and infrared image data is actually used for collecting pictures of a person drowning scene by deploying visible light and infrared photographic equipment, and on the data set, a neural network structure for target detection is designed to establish a two-classification deep learning model suitable for a ship navigation environment.
The convolution neural network step realizes the feature extraction of the fused image, outputs a target detection result in a specific layer after multi-layer convolution and activation function, and feeds the result back to the control alarm module;
inputting real-time image data into a deep learning model, calculating whether a current image contains a person falling into water or not according to the network weight obtained by training, and if the current image contains the person falling into water, marking;
the convolutional neural network model takes a residual error network as a main body, and a neural network layer comprises a convolutional layer and a pooling layer;
the deep learning model is a convolutional neural network model;
when the deep learning model is trained by using the visible light image data and the infrared image data, the visible light image data and the infrared image data comprise image data of people who fall into water and image data of people who do not fall into water, and the quantity of the image data of the people who fall into water is equivalent to that of the image data of the people who do not fall into water.
In the present embodiment, the convolutional neural network is mainly composed of a residual network, and is composed of a plurality of network layers, including convolutional layers, pooling layers, and the like, and adopts a full convolutional layer structure, and the size of an input image is variable;
controlling and alarming: the judgment result of the convolutional neural network module is fed back to the control alarm module, when a dangerous case occurs, the background alarm is controlled to be turned on or off, and the water falling information is stored so as to facilitate history tracing;
specifically, the step of controlling the alarm includes: the alarm device comprises an alarm device unit and an alarm information storage unit;
when the control alarm module obtains the dangerous case information, the alarm unit starts an alarm; reminding workers to intervene, storing the information falling into water by the alarm information storage unit, storing the information falling into water into the SQLITE database, calling real-time images of the image acquisition module by the workers in time for intervention, calling playback video segments of the video storage module, and reasonably deploying rescue actions after detailed judgment. The historical data can be conveniently traced;
the drowning information comprises a drowning time and a drowning position.
An image display step: and the image display module displays the output result of the convolutional neural network module and the historical data of the image storage module.
Specifically, the image displaying step includes: displaying different image data according to different output results of the convolutional neural network module, and when an emergency occurs, displaying data of an emergency area and marking people falling into water; when no dangerous case occurs, displaying real-time images of each monitoring area;
meanwhile, a history playback function is realized, the image display module displays data of the image storage module, and a worker can read videos from the image storage module at a selected time period.
It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are merely illustrative of the principles of the invention, but that various changes and modifications may be made without departing from the spirit and scope of the invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.
Those skilled in the art will appreciate that, in addition to implementing the systems, apparatus, and various modules thereof provided by the present invention in purely computer readable program code, the same procedures can be implemented entirely by logically programming method steps such that the systems, apparatus, and various modules thereof are provided in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers and the like. Therefore, the system, the device and the modules thereof provided by the present invention can be considered as a hardware component, and the modules included in the system, the device and the modules thereof for implementing various programs can also be considered as structures in the hardware component; modules for performing various functions may also be considered to be both software programs for performing the methods and structures within hardware components.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes or modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention. The embodiments and features of the embodiments of the present application may be combined with each other arbitrarily without conflict.