WO2023065497A1 - Low-illumination obstacle detection method and system, and terminal and storage medium - Google Patents
Low-illumination obstacle detection method and system, and terminal and storage medium Download PDFInfo
- Publication number
- WO2023065497A1 WO2023065497A1 PCT/CN2021/137601 CN2021137601W WO2023065497A1 WO 2023065497 A1 WO2023065497 A1 WO 2023065497A1 CN 2021137601 W CN2021137601 W CN 2021137601W WO 2023065497 A1 WO2023065497 A1 WO 2023065497A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- low
- light
- image
- illumination
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Definitions
- the present application belongs to the technical field of natural image processing, and in particular relates to a low-light obstacle detection method, system, terminal and storage medium.
- an indoor visual assistance system based on Faster-RCNN, which detects daily objects such as fans, chairs, tables, and dehumidifiers in an indoor environment, and notifies users of their current location and plans movement routes through smart devices.
- Tapu et al developed a wearable device that can detect and alert obstacles with the help of YOLOv3, and designed a smartphone-based object detection and classification scheme to detect objects (cars, bicycles, people, etc.) in an outdoor environment to help visually impaired people; SonayDuman et al. also invented a portable guide system based on YOLO to help visually impaired people perceive objects and people around them and accurately estimate the distance between them.
- the existing blind-guiding equipment based on target detection can complete the tasks of identifying obstacles, recognizing faces of people, or guiding the route to the destination, etc.
- most of the existing blind-guiding equipment or technologies are only applicable to normal It is difficult to meet the needs of blind guides in low-light scenes.
- the present application provides a low-light obstacle detection method, system, terminal and storage medium, aiming to solve one of the above-mentioned technical problems in the prior art at least to a certain extent.
- a low-light obstacle detection method comprising:
- Obstacle recognition is performed on the low-light image to be detected through the trained target detection model.
- the conversion of the normal illumination sample image into the low illumination sample image by using the cyclic confrontation network includes:
- the cyclic confrontation network includes a generator and a discriminator, the normal illumination image is converted by the generator, and whether the converted image meets the requirements of the low-light image is judged by the discriminator, and if not, the converted image is discarded; if If the low-light image requirements are met, put the transformed image into the low-light image dataset.
- the enhancement processing of the low-light sample image through the illumination enhancement network includes:
- the illumination enhancement network includes a CNN network, the input of the CNN network is a low-light image data set, and during network training, the CNN network utilizes the relationship between the low-light image histogram and the normal illumination image histogram curve and A loss function is designed for data distribution, a self-supervised learning network is formed, and a mapping matrix for characterizing illumination enhancement features is output.
- the enhancement processing of the low-light sample image through the illumination enhancement network further includes:
- the illumination enhancement network also includes a residual connection adjustment module, the residual connection adjustment module includes four convolutional layers with residual connections; the input of the residual connection adjustment module is a low-light image data set, low After the low-light images in the light image data set are processed by four convolutional layers, three grayscale parameter matrices are obtained, and the mapping matrix output by the CNN network is iteratively adjusted through the three grayscale parameter matrices to obtain each low-light image The image corresponds to the resized image.
- the residual connection adjustment module includes four convolutional layers with residual connections; the input of the residual connection adjustment module is a low-light image data set, low After the low-light images in the light image data set are processed by four convolutional layers, three grayscale parameter matrices are obtained, and the mapping matrix output by the CNN network is iteratively adjusted through the three grayscale parameter matrices to obtain each low-light image The image corresponds to the resized image.
- the enhancement processing of the low-light sample image through the illumination enhancement network further includes:
- Each low-light image is separately added to the corresponding adjusted image to generate a low-light enhanced image.
- the inputting the low-light enhanced image into the target detection model for training includes:
- the target detection algorithm of the target detection model includes YOLO algorithm or Fast R-CNN detection algorithm.
- a low-light obstacle detection system including:
- Image conversion module used to convert the normal illumination sample image into a low illumination sample image by using the cyclic confrontation network
- Image enhancement module used to enhance the low-light sample image through the light-enhancing network to obtain a low-light enhanced image
- Model training module used to input the low-light enhanced image into the target detection model for training to obtain a trained target detection model
- Image recognition module used for performing obstacle recognition on the low-light image to be detected through the trained target detection model.
- a terminal includes a processor and a memory coupled to the processor, wherein,
- the memory stores program instructions for implementing the low-light obstacle detection method
- the processor is configured to execute the program instructions stored in the memory to control low-light obstacle detection.
- Yet another technical solution adopted in the embodiment of the present application is: a storage medium storing program instructions executable by a processor, the program instructions being used to execute the low-light obstacle detection method.
- the beneficial effects produced by the embodiments of the present application are that the low-light obstacle detection method, system, terminal and storage medium of the embodiments of the present application enhance low-light images by introducing a self-supervised low-light enhancement algorithm, Effectively combine the low-light enhancement with the target detection algorithm to improve the robustness and accuracy of the target detection algorithm for low-light target detection, and solve the problem of poor detection effect of using ordinary target detection algorithms for obstacle detection in low-light scenarios question. And adding a residual connection module can make better use of the original low-light image information, avoid the structural damage of the light enhancement network to the low-light image, and further improve the target detection accuracy of the target detection algorithm in low-light conditions.
- Fig. 1 is the flow chart of the low-light obstacle detection method of the embodiment of the present application.
- Fig. 2 is a schematic diagram of the structure of the cyclic confrontation network of the embodiment of the present application.
- FIG. 3 is a schematic structural diagram of a low-light obstacle detection system according to an embodiment of the present application.
- FIG. 4 is a schematic structural diagram of a terminal according to an embodiment of the present application.
- FIG. 5 is a schematic structural diagram of a storage medium according to an embodiment of the present application.
- FIG. 1 is a flow chart of a low-light obstacle detection method according to an embodiment of the present application.
- the low-light obstacle detection method of the embodiment of the present application includes the following steps:
- the normal illumination sample image may be obtained from a public dataset of normal illumination images.
- the embodiment of the present application uses a cycle confrontation network (CycleGAN) to convert the normal illumination sample image into a low illumination sample image.
- CycleGAN cycle confrontation network
- FIG. 2 it is a schematic diagram of the structure of the cyclic confrontation network in the embodiment of the present application.
- the cyclic confrontation network includes a generator and a discriminator.
- the A-style data set represents normal-light images
- the B-style data set represents low-light images.
- the normal-light images in the A-style data set are converted into low-light images through the generator G A. Input the converted image into the discriminator D A to judge whether it conforms to the B style, if not, discard the image, if it conforms to the B style, then put the image into the B style data set, so as to convert the normal light sample image into low light sample image.
- the image in the B-style data set is converted into a normal illumination image through the generator G B , and the converted image is input into the discriminator D B to judge whether it conforms to the A style. If it does not conform, the image is discarded. If it conforms to the A style, Then put the image into the A-style dataset to convert the low-light sample image into a normal-light sample image.
- a low-light data set for training the target detection network can be generated to provide sufficient training data support for the target detection network.
- S30 Input the low-light image dataset into the illumination enhancement network for enhancement processing, and output the low-light enhancement image dataset through the illumination enhancement network;
- the illumination enhancement network includes a CNN (convolutional neural network) network and a residual connection adjustment module.
- the input of the CNN network is a low-light image data set.
- the CNN network uses the low-light image histogram and The relationship between the histogram curves of the normal illumination image and the data distribution design a loss function to form a self-supervised learning network, and output a mapping matrix for representing illumination enhancement features.
- the input of the residual connection adjustment module is a low-light image dataset, including four convolutional layers with residual connections.
- three grayscale parameter matrices are obtained. Through three The grayscale parameter matrix performs three iterative adjustments on the mapping matrix output by the CNN network to obtain the adjusted image corresponding to each low-light image.
- each low-light image is added to the corresponding adjusted image to generate a low-light enhanced image.
- the residual connection adjustment module in the embodiment of the present application can effectively fine-tune the weight of the illumination enhancement network during network training, and can make better use of the original image information, avoiding the structure of the illumination enhancement network on low-light images. on the destruction.
- S40 Input the low-light enhanced image data set into the target detection model for training, and output the target category and position information in the low-light enhanced image through the target detection model;
- the target detection model can use the YOLO algorithm or the Fast R-CNN detection algorithm.
- S50 Input the low-light image to be detected into the trained target detection model, and perform obstacle recognition on the low-light image through the target detection model.
- the low-light obstacle detection method of the embodiment of the present application introduces a self-supervised low-light enhancement algorithm to enhance the low-light image, effectively combines the low-light enhancement with the target detection algorithm, and improves the accuracy of the target detection algorithm for low-light targets.
- the robustness and precision of the detection solve the problem of poor detection effect of using ordinary target detection algorithms for obstacle detection in low-light scenarios.
- adding a residual connection module can make better use of the original low-light image information, avoid the structural damage of the light enhancement network to the low-light image, and further improve the target detection accuracy of the target detection algorithm in low-light conditions.
- FIG. 3 is a schematic structural diagram of a low-light obstacle detection system according to an embodiment of the present application.
- the low-light obstacle detection system 40 of the embodiment of the present application includes:
- Image conversion module 41 used to convert the normal illumination sample image into a low illumination sample image by using the cyclic confrontation network
- Image enhancement module 42 used to enhance the low-light sample image through the light-enhancing network to obtain a low-light enhanced image
- Model training module 43 used for inputting low-light enhanced images into the target detection model for training to obtain a trained target detection model
- Image recognition module 44 used for performing obstacle recognition on the low-light image to be detected through the trained target detection model.
- FIG. 4 is a schematic diagram of a terminal structure in an embodiment of the present application.
- the terminal 50 includes a processor 51 and a memory 52 coupled to the processor 51 .
- the memory 52 stores program instructions for realizing the above-mentioned low-light obstacle detection method.
- the processor 51 is used to execute the program instructions stored in the memory 52 to control low-light obstacle detection.
- the processor 51 can also be referred to as a CPU (Central Processing Unit, central processing unit).
- the processor 51 may be an integrated circuit chip with signal processing capability.
- the processor 51 can also be a general-purpose processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components .
- a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
- FIG. 5 is a schematic structural diagram of a storage medium according to an embodiment of the present application.
- the storage medium in the embodiment of the present application stores a program file 61 capable of implementing all the above-mentioned methods, wherein the program file 61 can be stored in the above-mentioned storage medium in the form of a software product, and includes several instructions to make a computer device (which can It is a personal computer, a server, or a network device, etc.) or a processor (processor) that executes all or part of the steps of the methods in various embodiments of the present application.
- a computer device which can It is a personal computer, a server, or a network device, etc.
- processor processor
- the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk, and other media that can store program codes. , or terminal devices such as computers, servers, mobile phones, and tablets.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
Description
本申请属于自然图像处理技术领域,特别涉及一种低光照障碍物检测方法、系统、终端以及存储介质。The present application belongs to the technical field of natural image processing, and in particular relates to a low-light obstacle detection method, system, terminal and storage medium.
根据世界卫生组织统计,全世界有近2.53亿人群患有视力障碍,其中有3600万完全丧失了视觉,与此同时人口的增长和老龄化的加剧也将使视力障碍人群的规模进一步加剧,因此低视力导盲设备也有着不容忽视的巨大需求。According to the statistics of the World Health Organization, nearly 253 million people around the world suffer from visual impairment, and 36 million of them have completely lost their vision. Low vision guide equipment also has a huge demand that cannot be ignored.
传统的导盲手段包括导盲棒、导盲犬、盲文、盲道等,但这些导盲手段的安全性并不完全可靠,且价格昂贵、适用范围狭窄。近年来,随着深度学习和其在计算机视觉领域的突破性发展,导盲设备的发展迎来了新的生机。目前基于深度学习目标检测的主流框架可以大体上分为以Fast R-CNN为代表的双阶段目标检测算法和以YOLO系列、SSD等为代表的单阶段目标检测算法两个大的技术方向,这些深度学习算法相较传统算法效果有了突破性的提升,给目标检测技术带来新的研究热潮。例如Lin等人开发了一种基于Faster-RCNN的室内视觉辅助系统,在室内环境中对风扇,椅子,桌子和除湿机等日常物品进行检测,并通过智能设备通知用户当前位置并规划移动路线。Tapu等借助YOLOv3开发了能检测并警报障碍物的可穿戴设备,设计了基于智能手机的对象检测和分类方案,以在户外环境中对物体(汽车,自行车,人等)进行检测来帮助视力障碍的人;SonayDuman等也发明一种基于YOLO的便携式导盲系统,帮助视障人士感知周围的物体和人,并精确估计他们之间的距离。Traditional blind-guiding means include guide sticks, guide dogs, Braille, blind trails, etc., but the safety of these blind-guiding means is not completely reliable, and they are expensive and have a narrow scope of application. In recent years, with deep learning and its breakthrough development in the field of computer vision, the development of blind guide equipment has ushered in new vitality. At present, the mainstream framework of target detection based on deep learning can be roughly divided into two major technical directions: the two-stage target detection algorithm represented by Fast R-CNN and the single-stage target detection algorithm represented by YOLO series and SSD. Compared with the traditional algorithm, the deep learning algorithm has a breakthrough improvement, which brings a new research boom to the target detection technology. For example, Lin et al. developed an indoor visual assistance system based on Faster-RCNN, which detects daily objects such as fans, chairs, tables, and dehumidifiers in an indoor environment, and notifies users of their current location and plans movement routes through smart devices. Tapu et al developed a wearable device that can detect and alert obstacles with the help of YOLOv3, and designed a smartphone-based object detection and classification scheme to detect objects (cars, bicycles, people, etc.) in an outdoor environment to help visually impaired people; SonayDuman et al. also invented a portable guide system based on YOLO to help visually impaired people perceive objects and people around them and accurately estimate the distance between them.
上述中,现有基于目标检测的导盲设备虽然可以完成障碍物的识别、人物面部的识别或指引到达目的地的路线等等的任务,但是现有的导盲设备或技术大多只适用于正常的光照条件,难以应对低光照场景下的导盲需求。Among the above, although the existing blind-guiding equipment based on target detection can complete the tasks of identifying obstacles, recognizing faces of people, or guiding the route to the destination, etc., most of the existing blind-guiding equipment or technologies are only applicable to normal It is difficult to meet the needs of blind guides in low-light scenes.
本申请提供了一种低光照障碍物检测方法、系统、终端以及存储介质,旨在至少在一定程度上解决现有技术中的上述技术问题之一。The present application provides a low-light obstacle detection method, system, terminal and storage medium, aiming to solve one of the above-mentioned technical problems in the prior art at least to a certain extent.
为了解决上述问题,本申请提供了如下技术方案:In order to solve the above problems, the application provides the following technical solutions:
一种低光照障碍物检测方法,包括:A low-light obstacle detection method, comprising:
利用循环对抗网络将正常光照样本图像转换为低光照样本图像;Convert normal lighting sample images to low lighting sample images using a recurrent adversarial network;
通过光照增强网络对所述低光照样本图像进行增强处理,得到低光照增强图像;performing enhancement processing on the low-light sample image through a light-enhancing network to obtain a low-light enhanced image;
将所述低光照增强图像输入目标检测模型进行训练,得到训练好的目标检测模型;Inputting the low-light enhanced image into a target detection model for training to obtain a trained target detection model;
通过所述训练好的目标检测模型对待检测的低光照图像进行障碍物识别。Obstacle recognition is performed on the low-light image to be detected through the trained target detection model.
本申请实施例采取的技术方案还包括:所述利用循环对抗网络将正常光照样本图像转换为低光照样本图像包括:The technical solution adopted in the embodiment of the present application also includes: the conversion of the normal illumination sample image into the low illumination sample image by using the cyclic confrontation network includes:
所述循环对抗网络包括生成器和判别器,通过所述生成器对正常光照图像进行转换,并通过所述判别器判别转换图像是否符合低光照图像要求,如果不符合则丢弃该转换图像;如果符合低光照图像要求,则将该转换图像放入低光照图像数据集。The cyclic confrontation network includes a generator and a discriminator, the normal illumination image is converted by the generator, and whether the converted image meets the requirements of the low-light image is judged by the discriminator, and if not, the converted image is discarded; if If the low-light image requirements are met, put the transformed image into the low-light image dataset.
本申请实施例采取的技术方案还包括:所述通过光照增强网络对所述低光照样本图像进行增强处理包括:The technical solution adopted in the embodiment of the present application further includes: the enhancement processing of the low-light sample image through the illumination enhancement network includes:
所述光照增强网络包括一个CNN网络,所述CNN网络的输入为低光照图像数据集,在网络训练时,所述CNN网络利用低光照图像直方图与正常光照图像直方图曲线之间的关系和数据分布设计损失函数,构成自监督学习网络,并输出用于表征光照增强特征的映射矩阵。The illumination enhancement network includes a CNN network, the input of the CNN network is a low-light image data set, and during network training, the CNN network utilizes the relationship between the low-light image histogram and the normal illumination image histogram curve and A loss function is designed for data distribution, a self-supervised learning network is formed, and a mapping matrix for characterizing illumination enhancement features is output.
本申请实施例采取的技术方案还包括:所述通过光照增强网络对所述低光照样本图像进行增强处理还包括:The technical solution adopted in the embodiment of the present application further includes: the enhancement processing of the low-light sample image through the illumination enhancement network further includes:
所述光照增强网络还包括残差连接调整模块,所述残差连接调整模块包括带有残差连接的四个卷积层;所述残差连接调整模块的输入为低光照图像数据集,低光照图像数据集中的低光照图像经过四个卷积层处理后,得到三个灰度参数矩阵,通过所述三个灰度参数矩阵对CNN网络输出的映射矩阵进行迭代调整,得到每幅低光照图像对应的调整图像。The illumination enhancement network also includes a residual connection adjustment module, the residual connection adjustment module includes four convolutional layers with residual connections; the input of the residual connection adjustment module is a low-light image data set, low After the low-light images in the light image data set are processed by four convolutional layers, three grayscale parameter matrices are obtained, and the mapping matrix output by the CNN network is iteratively adjusted through the three grayscale parameter matrices to obtain each low-light image The image corresponds to the resized image.
本申请实施例采取的技术方案还包括:所述通过光照增强网络对所述低光照样本图像进行增强处理还包括:The technical solution adopted in the embodiment of the present application further includes: the enhancement processing of the low-light sample image through the illumination enhancement network further includes:
分别将每幅低光照图像与对应的调整图像相加,生成低光照增强图像。Each low-light image is separately added to the corresponding adjusted image to generate a low-light enhanced image.
本申请实施例采取的技术方案还包括:所述将所述低光照增强图像输入目标检测模型进行训练包括:The technical solution adopted in the embodiment of the present application further includes: the inputting the low-light enhanced image into the target detection model for training includes:
所述目标检测模型的目标检测算法包括YOLO算法或Fast R-CNN检测算法。The target detection algorithm of the target detection model includes YOLO algorithm or Fast R-CNN detection algorithm.
本申请实施例采取的另一技术方案为:一种低光照障碍物检测系统,包括:Another technical solution adopted in the embodiment of the present application is: a low-light obstacle detection system, including:
图像转换模块:用于利用循环对抗网络将正常光照样本图像转换为低光照样本图像;Image conversion module: used to convert the normal illumination sample image into a low illumination sample image by using the cyclic confrontation network;
图像增强模块:用于通过光照增强网络对所述低光照样本图像进行增强处理,得到低光照增强图像;Image enhancement module: used to enhance the low-light sample image through the light-enhancing network to obtain a low-light enhanced image;
模型训练模块:用于将所述低光照增强图像输入目标检测模型进行训练,得到训练好的目标检测模型;Model training module: used to input the low-light enhanced image into the target detection model for training to obtain a trained target detection model;
图像识别模块:用于通过所述训练好的目标检测模型对待检测的低光照图像进行障碍物识别。Image recognition module: used for performing obstacle recognition on the low-light image to be detected through the trained target detection model.
本申请实施例采取的又一技术方案为:一种终端,所述终端包括处理器、与所述处理器耦接的存储器,其中, Another technical solution adopted by the embodiment of the present application is: a terminal, the terminal includes a processor and a memory coupled to the processor, wherein,
所述存储器存储有用于实现所述低光照障碍物检测方法的程序指令;The memory stores program instructions for implementing the low-light obstacle detection method;
所述处理器用于执行所述存储器存储的所述程序指令以控制低光照障碍物检测。The processor is configured to execute the program instructions stored in the memory to control low-light obstacle detection.
本申请实施例采取的又一技术方案为:一种存储介质,存储有处理器可运行的程序指令,所述程序指令用于执行所述低光照障碍物检测方法。Yet another technical solution adopted in the embodiment of the present application is: a storage medium storing program instructions executable by a processor, the program instructions being used to execute the low-light obstacle detection method.
相对于现有技术,本申请实施例产生的有益效果在于:本申请实施例的低光照障碍物检测方法、系统、终端以及存储介质通过引入自监督低光照增强算法对低光照图像进行增强处理,将低光照增强与目标检测算法进行有效结合,提高目标检测算法对低光照目标检测的鲁棒性和精度,解决了在低光照情景下使用普通目标检测算法进行障碍物检测存在的检测效果差的问题。并加入残差连接模块,能够更好的利用原始低光照图像信息,避免光照增强网络对低光照图像产生结构上的破坏,进一步提高目标检测算法在低光照情况下的目标检测精度。Compared with the prior art, the beneficial effects produced by the embodiments of the present application are that the low-light obstacle detection method, system, terminal and storage medium of the embodiments of the present application enhance low-light images by introducing a self-supervised low-light enhancement algorithm, Effectively combine the low-light enhancement with the target detection algorithm to improve the robustness and accuracy of the target detection algorithm for low-light target detection, and solve the problem of poor detection effect of using ordinary target detection algorithms for obstacle detection in low-light scenarios question. And adding a residual connection module can make better use of the original low-light image information, avoid the structural damage of the light enhancement network to the low-light image, and further improve the target detection accuracy of the target detection algorithm in low-light conditions.
图1是本申请实施例的低光照障碍物检测方法的流程图;Fig. 1 is the flow chart of the low-light obstacle detection method of the embodiment of the present application;
图2为本申请实施例的循环对抗网络结构示意图;Fig. 2 is a schematic diagram of the structure of the cyclic confrontation network of the embodiment of the present application;
图3为本申请实施例的低光照障碍物检测系统结构示意图;FIG. 3 is a schematic structural diagram of a low-light obstacle detection system according to an embodiment of the present application;
图4为本申请实施例的终端结构示意图;FIG. 4 is a schematic structural diagram of a terminal according to an embodiment of the present application;
图5为本申请实施例的存储介质的结构示意图。FIG. 5 is a schematic structural diagram of a storage medium according to an embodiment of the present application.
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solution and advantages of the present application clearer, the present application will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present application, not to limit the present application.
请参阅图1,是本申请实施例的低光照障碍物检测方法的流程图。本申请实施例的低光照障碍物检测方法包括以下步骤:Please refer to FIG. 1 , which is a flow chart of a low-light obstacle detection method according to an embodiment of the present application. The low-light obstacle detection method of the embodiment of the present application includes the following steps:
S10:获取一定数量的正常光照样本图像;S10: Obtain a certain number of normal illumination sample images;
本步骤中,可通常正常光照图像的公共数据集获取正常光照样本图像。In this step, the normal illumination sample image may be obtained from a public dataset of normal illumination images.
S20:利用循环对抗网络将正常光照样本图像转换为低光照样本图像,生成用于训练目标检测网络的低光照数据集;S20: Using the recurrent confrontation network to convert the normal illumination sample image into a low illumination sample image, and generate a low illumination data set for training the target detection network;
本步骤中,现有的光照图像公共数据集大多都是正常光照图像,低光照图像数据集提供的数据量十分有限(例如Exdark数据集)。为了保证能够对目标检测网络进行充分训练,本申请实施例利用循环对抗网络(CycleGAN)将正常光照样本图像转换为低光照样本图像。具体如图2所示,为本申请实施例的循环对抗网络结构示意图。其中,循环对抗网络包括生成器和判别器,A风格数据集代表正常光照图像,B风格数据集代表低光照图像,通过生成器G A将A风格数据集中的正常光照图像转换为低光照图像,将转换后的图像输入判别器D A判别是否符合B风格,如果不符合则丢弃该图像,如果符合B风格,则将该图像放入B风格数据集,从而将正常光照样本图像转换为低光照样本图像。同理,通过生成器G B将B风格数据集中的图像转换为正常光照图像,将转换后的图像输入判别器D B判别是否符合A风格,如果不符合则丢弃该图像,如果符合A风格,则将该图像放入A风格数据集,从而将低光照样本图像转换为正常光照样本图像。如此循环,在网络训练结束后,即可生成用于训练目标检测网络的低光照数据集,为目标检测网络提供充分的训练数据支撑。 In this step, most of the existing public datasets of illumination images are normal illumination images, and the amount of data provided by low-light image datasets is very limited (such as the Exdark dataset). In order to ensure that the target detection network can be fully trained, the embodiment of the present application uses a cycle confrontation network (CycleGAN) to convert the normal illumination sample image into a low illumination sample image. Specifically, as shown in FIG. 2 , it is a schematic diagram of the structure of the cyclic confrontation network in the embodiment of the present application. Among them, the cyclic confrontation network includes a generator and a discriminator. The A-style data set represents normal-light images, and the B-style data set represents low-light images. The normal-light images in the A-style data set are converted into low-light images through the generator G A. Input the converted image into the discriminator D A to judge whether it conforms to the B style, if not, discard the image, if it conforms to the B style, then put the image into the B style data set, so as to convert the normal light sample image into low light sample image. In the same way, the image in the B-style data set is converted into a normal illumination image through the generator G B , and the converted image is input into the discriminator D B to judge whether it conforms to the A style. If it does not conform, the image is discarded. If it conforms to the A style, Then put the image into the A-style dataset to convert the low-light sample image into a normal-light sample image. In such a cycle, after the network training is completed, a low-light data set for training the target detection network can be generated to provide sufficient training data support for the target detection network.
S30:将低光照图像数据集输入光照增强网络进行增强处理,通过光照增强网络输出低光照增强图像数据集;S30: Input the low-light image dataset into the illumination enhancement network for enhancement processing, and output the low-light enhancement image dataset through the illumination enhancement network;
本步骤中,光照增强网络包括一个CNN(卷积神经网络)网络和一个残差连接调整模块, CNN网络的输入为低光照图像数据集,在网络训练时,CNN网络利用低光照图像直方图与正常光照图像直方图曲线之间的关系和数据分布设计损失函数,构成自监督学习网络,并输出用于表征光照增强特征的映射矩阵。残差连接调整模块的输入为低光照图像数据集,包括带有残差连接的四个卷积层,低光照图像经过四个卷积层处理后,得到三个灰度参数矩阵,通过三个灰度参数矩阵对CNN网络输出的映射矩阵进行三次迭代调整,得到每幅低光照图像对应的调整图像,最后,分别将每幅低光照图像与对应的调整图像相加,生成低光照增强图像。In this step, the illumination enhancement network includes a CNN (convolutional neural network) network and a residual connection adjustment module. The input of the CNN network is a low-light image data set. During network training, the CNN network uses the low-light image histogram and The relationship between the histogram curves of the normal illumination image and the data distribution design a loss function to form a self-supervised learning network, and output a mapping matrix for representing illumination enhancement features. The input of the residual connection adjustment module is a low-light image dataset, including four convolutional layers with residual connections. After the low-light image is processed by four convolutional layers, three grayscale parameter matrices are obtained. Through three The grayscale parameter matrix performs three iterative adjustments on the mapping matrix output by the CNN network to obtain the adjusted image corresponding to each low-light image. Finally, each low-light image is added to the corresponding adjusted image to generate a low-light enhanced image.
上述中,本申请实施例中的残差连接调整模块可以在网络训练时有效的对光照增强网络的权重进行微调,并能够更好的利用原始图像信息,避免光照增强网络对低光照图像产生结构上的破坏。In the above, the residual connection adjustment module in the embodiment of the present application can effectively fine-tune the weight of the illumination enhancement network during network training, and can make better use of the original image information, avoiding the structure of the illumination enhancement network on low-light images. on the destruction.
S40:将低光照增强图像数据集输入目标检测模型进行训练,通过目标检测模型输出低光照增强图像中的目标类别及位置信息;S40: Input the low-light enhanced image data set into the target detection model for training, and output the target category and position information in the low-light enhanced image through the target detection model;
本步骤中,目标检测模型可以使用YOLO算法或者Fast R-CNN检测算法等。In this step, the target detection model can use the YOLO algorithm or the Fast R-CNN detection algorithm.
S50:将待检测的低光照图像输入训练好的目标检测模型,通过目标检测模型对低光照图像进行障碍物识别。S50: Input the low-light image to be detected into the trained target detection model, and perform obstacle recognition on the low-light image through the target detection model.
基于上述,本申请实施例的低光照障碍物检测方法通过引入自监督低光照增强算法对低光照图像进行增强处理,将低光照增强与目标检测算法进行有效结合,提高目标检测算法对低光照目标检测的鲁棒性和精度,解决了在低光照情景下使用普通目标检测算法进行障碍物检测存在的检测效果差的问题。并加入残差连接模块,能够更好的利用原始低光照图像信息,避免光照增强网络对低光照图像产生结构上的破坏,进一步提高目标检测算法在低光照情况下的目标检测精度。Based on the above, the low-light obstacle detection method of the embodiment of the present application introduces a self-supervised low-light enhancement algorithm to enhance the low-light image, effectively combines the low-light enhancement with the target detection algorithm, and improves the accuracy of the target detection algorithm for low-light targets. The robustness and precision of the detection solve the problem of poor detection effect of using ordinary target detection algorithms for obstacle detection in low-light scenarios. And adding a residual connection module can make better use of the original low-light image information, avoid the structural damage of the light enhancement network to the low-light image, and further improve the target detection accuracy of the target detection algorithm in low-light conditions.
请参阅图3,为本申请实施例的低光照障碍物检测系统结构示意图。本申请实施例的低光照障碍物检测系统40包括:Please refer to FIG. 3 , which is a schematic structural diagram of a low-light obstacle detection system according to an embodiment of the present application. The low-light obstacle detection system 40 of the embodiment of the present application includes:
图像转换模块41:用于利用循环对抗网络将正常光照样本图像转换为低光照样本图像;Image conversion module 41: used to convert the normal illumination sample image into a low illumination sample image by using the cyclic confrontation network;
图像增强模块42:用于通过光照增强网络对低光照样本图像进行增强处理,得到低光照增强图像;Image enhancement module 42: used to enhance the low-light sample image through the light-enhancing network to obtain a low-light enhanced image;
模型训练模块43:用于将低光照增强图像输入目标检测模型进行训练,得到训练好的目标检测模型;Model training module 43: used for inputting low-light enhanced images into the target detection model for training to obtain a trained target detection model;
图像识别模块44:用于通过训练好的目标检测模型对待检测的低光照图像进行障碍物识别。Image recognition module 44: used for performing obstacle recognition on the low-light image to be detected through the trained target detection model.
请参阅图4,为本申请实施例的终端结构示意图。该终端50包括处理器51、与处理器51耦接的存储器52。Please refer to FIG. 4 , which is a schematic diagram of a terminal structure in an embodiment of the present application. The terminal 50 includes a processor 51 and a memory 52 coupled to the processor 51 .
存储器52存储有用于实现上述低光照障碍物检测方法的程序指令。The memory 52 stores program instructions for realizing the above-mentioned low-light obstacle detection method.
处理器51用于执行存储器52存储的程序指令以控制低光照障碍物检测。The processor 51 is used to execute the program instructions stored in the memory 52 to control low-light obstacle detection.
其中,处理器51还可以称为CPU(Central Processing Unit,中央处理单元)。处理器51可能是一种集成电路芯片,具有信号的处理能力。处理器51还可以是通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现成可编程门阵列(FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。Wherein, the processor 51 can also be referred to as a CPU (Central Processing Unit, central processing unit). The processor 51 may be an integrated circuit chip with signal processing capability. The processor 51 can also be a general-purpose processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components . A general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
请参阅图5,为本申请实施例的存储介质的结构示意图。本申请实施例的存储介质存储有能够实现上述所有方法的程序文件61,其中,该程序文件61可以以软件产品的形式存储在上述存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器(processor)执行本申请各个实施方式方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质,或者是计算机、服务器、手机、平板等终端设备。Please refer to FIG. 5 , which is a schematic structural diagram of a storage medium according to an embodiment of the present application. The storage medium in the embodiment of the present application stores a program file 61 capable of implementing all the above-mentioned methods, wherein the program file 61 can be stored in the above-mentioned storage medium in the form of a software product, and includes several instructions to make a computer device (which can It is a personal computer, a server, or a network device, etc.) or a processor (processor) that executes all or part of the steps of the methods in various embodiments of the present application. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk, and other media that can store program codes. , or terminal devices such as computers, servers, mobile phones, and tablets.
对所公开的实施例的上述说明,使本领域专业技术人员能够实现或使用本申请。对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见的,本申请中所定义的一般原理可以在不脱离本申请的精神或范围的情况下,在其它实施例中实现。因此,本申请将不会被限制于本申请所示的这些实施例,而是要符合与本申请所公开的原理和新颖特点相一致的最宽的范围。The above description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the general principles defined in this application may be implemented in other embodiments without departing from the spirit or scope of the application. Therefore, the present application will not be limited to the embodiments shown in the present application, but is to be accorded the widest scope consistent with the principles and novel features disclosed in the present application.
Claims (9)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202111235613.X | 2021-10-22 | ||
| CN202111235613.XA CN114065838B (en) | 2021-10-22 | 2021-10-22 | A low-light obstacle detection method, system, terminal and storage medium |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2023065497A1 true WO2023065497A1 (en) | 2023-04-27 |
Family
ID=80235298
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2021/137601 Ceased WO2023065497A1 (en) | 2021-10-22 | 2021-12-13 | Low-illumination obstacle detection method and system, and terminal and storage medium |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN114065838B (en) |
| WO (1) | WO2023065497A1 (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117036853A (en) * | 2023-08-01 | 2023-11-10 | 智洋创新科技股份有限公司 | Image enhancement and target detection accuracy rate improvement method based on joint training |
| CN117893880A (en) * | 2024-01-25 | 2024-04-16 | 西南科技大学 | An object detection method based on adaptive feature learning in low-light images |
| CN119399448A (en) * | 2024-11-18 | 2025-02-07 | 南京邮电大学 | A low-illumination target detection method with multi-scale receptive field under adaptive enhancement |
| CN119904730A (en) * | 2024-12-30 | 2025-04-29 | 浙江大学 | A cross-variety target detection method, device, equipment and storage medium |
Families Citing this family (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114663950B (en) * | 2022-03-28 | 2025-08-15 | 深圳市赛为智能股份有限公司 | Low-illumination face detection method and device, computer equipment and storage medium |
| CN114663951B (en) * | 2022-03-28 | 2025-08-01 | 深圳市赛为智能股份有限公司 | Low-illumination face detection method and device, computer equipment and storage medium |
| CN115170415B (en) * | 2022-06-30 | 2025-04-22 | 广东工业大学 | A low-light image enhancement method, system and readable storage medium |
| CN115620090A (en) * | 2022-11-07 | 2023-01-17 | 中电科新型智慧城市研究院有限公司 | Model training method, low-illumination target re-identification method and device, and terminal equipment |
| CN116152923B (en) * | 2023-01-13 | 2025-10-21 | 厦门市美亚柏科信息股份有限公司 | A method and system for target re-identification at night |
| CN119131464B (en) * | 2024-08-16 | 2025-10-24 | 中国科学院沈阳计算技术研究所有限公司 | A low-light target detection method based on two-layer optimization |
| CN119863386A (en) * | 2025-03-24 | 2025-04-22 | 浙江优众新材料科技有限公司 | Low-light image enhancement method based on semi-supervision and related equipment |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20200065992A1 (en) * | 2018-08-23 | 2020-02-27 | Samsung Electronics Co., Ltd. | Method and apparatus for recognizing image and method and apparatus for training recognition model based on data augmentation |
| CN111402145A (en) * | 2020-02-17 | 2020-07-10 | 哈尔滨工业大学 | A self-supervised low-light image enhancement method based on deep learning |
| CN112560579A (en) * | 2020-11-20 | 2021-03-26 | 中国科学院深圳先进技术研究院 | Obstacle detection method based on artificial intelligence |
| CN112950561A (en) * | 2021-02-22 | 2021-06-11 | 中国地质大学(武汉) | Optical fiber end face defect detection method, device and storage medium |
| CN113052210A (en) * | 2021-03-11 | 2021-06-29 | 北京工业大学 | Fast low-illumination target detection method based on convolutional neural network |
| CN113392702A (en) * | 2021-05-10 | 2021-09-14 | 南京师范大学 | Target identification method based on self-adaptive image enhancement under low-light environment |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109886161B (en) * | 2019-01-30 | 2023-12-12 | 江南大学 | Road traffic identification recognition method based on likelihood clustering and convolutional neural network |
| CN110210401B (en) * | 2019-06-03 | 2022-10-21 | 多维协同人工智能技术研究院(重庆)有限公司 | Intelligent target detection method under weak light |
| CN111161178A (en) * | 2019-12-25 | 2020-05-15 | 湖南大学 | Single low-light image enhancement method based on generation type countermeasure network |
| CN111291885B (en) * | 2020-01-20 | 2023-06-09 | 北京百度网讯科技有限公司 | Near-infrared image generation method, generation network training method and device |
| CN111968044B (en) * | 2020-07-16 | 2023-08-11 | 中国科学院沈阳自动化研究所 | Low-illumination image enhancement method based on Retinex and deep learning |
| CN111798400B (en) * | 2020-07-20 | 2022-10-11 | 福州大学 | Reference-free low-light image enhancement method and system based on generative adversarial network |
| CN112614077B (en) * | 2020-12-30 | 2022-08-19 | 北京航空航天大学杭州创新研究院 | Unsupervised low-illumination image enhancement method based on generation countermeasure network |
| CN112766056B (en) * | 2020-12-30 | 2023-10-27 | 厦门大学 | Method and device for detecting lane lines in low-light environment based on deep neural network |
| CN113313657B (en) * | 2021-07-29 | 2021-12-21 | 北京航空航天大学杭州创新研究院 | An unsupervised learning method and system for low-light image enhancement |
-
2021
- 2021-10-22 CN CN202111235613.XA patent/CN114065838B/en active Active
- 2021-12-13 WO PCT/CN2021/137601 patent/WO2023065497A1/en not_active Ceased
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20200065992A1 (en) * | 2018-08-23 | 2020-02-27 | Samsung Electronics Co., Ltd. | Method and apparatus for recognizing image and method and apparatus for training recognition model based on data augmentation |
| CN111402145A (en) * | 2020-02-17 | 2020-07-10 | 哈尔滨工业大学 | A self-supervised low-light image enhancement method based on deep learning |
| CN112560579A (en) * | 2020-11-20 | 2021-03-26 | 中国科学院深圳先进技术研究院 | Obstacle detection method based on artificial intelligence |
| CN112950561A (en) * | 2021-02-22 | 2021-06-11 | 中国地质大学(武汉) | Optical fiber end face defect detection method, device and storage medium |
| CN113052210A (en) * | 2021-03-11 | 2021-06-29 | 北京工业大学 | Fast low-illumination target detection method based on convolutional neural network |
| CN113392702A (en) * | 2021-05-10 | 2021-09-14 | 南京师范大学 | Target identification method based on self-adaptive image enhancement under low-light environment |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117036853A (en) * | 2023-08-01 | 2023-11-10 | 智洋创新科技股份有限公司 | Image enhancement and target detection accuracy rate improvement method based on joint training |
| CN117893880A (en) * | 2024-01-25 | 2024-04-16 | 西南科技大学 | An object detection method based on adaptive feature learning in low-light images |
| CN119399448A (en) * | 2024-11-18 | 2025-02-07 | 南京邮电大学 | A low-illumination target detection method with multi-scale receptive field under adaptive enhancement |
| CN119904730A (en) * | 2024-12-30 | 2025-04-29 | 浙江大学 | A cross-variety target detection method, device, equipment and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| CN114065838A (en) | 2022-02-18 |
| CN114065838B (en) | 2023-07-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN114065838B (en) | A low-light obstacle detection method, system, terminal and storage medium | |
| Zhang et al. | Gender and smile classification using deep convolutional neural networks | |
| JP2022133378A (en) | Face biometric detection method, device, electronic device, and storage medium | |
| WO2022067668A1 (en) | Fire detection method and system based on video image target detection, and terminal and storage medium | |
| WO2021218899A1 (en) | Method for training facial recognition model, and method and apparatus for facial recognition | |
| CN118429389B (en) | Target tracking method and system based on multi-scale aggregated attention feature extraction network | |
| WO2019223254A1 (en) | Construction method for multi-scale lightweight face detection model and face detection method based on model | |
| CN115830392B (en) | Student behavior recognition method based on improved YOLOv5 | |
| WO2021023003A1 (en) | Face conversion model training method, storage medium, and terminal device | |
| CN113378770B (en) | Gesture recognition method, device, equipment and storage medium | |
| CN110633610A (en) | A student state detection algorithm based on YOLO | |
| US12080100B2 (en) | Face-aware person re-identification system | |
| CN111461212A (en) | A compression method for point cloud target detection model | |
| WO2019179295A1 (en) | Facial recognition method and device | |
| CN113239915B (en) | Classroom behavior identification method, device, equipment and storage medium | |
| CN111797733A (en) | An image-based behavior recognition method, device, device and storage medium | |
| CN114399655A (en) | Target detection method, system and storage medium | |
| CN108960047A (en) | Face De-weight method in video monitoring based on the secondary tree of depth | |
| WO2022178833A1 (en) | Target detection network training method, target detection method, and apparatus | |
| CN107330387B (en) | Pedestrian detection method based on image data | |
| CN114120456A (en) | Learning concentration detection method, computer equipment and readable medium | |
| Li et al. | Prototype-driven multi-feature generation for visible-infrared person re-identification | |
| CN115496776A (en) | Cutout method, cutout model training method and device, equipment, medium | |
| CN118366205A (en) | A lightweight face tracking method and system based on attention mechanism | |
| CN115953744A (en) | A vehicle recognition and tracking method based on deep learning |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 21961235 Country of ref document: EP Kind code of ref document: A1 |
|
| 32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10-04-2025) |