WO2019192121A1 - Dual-channel neural network model training and human face comparison method, and terminal and medium - Google Patents
Dual-channel neural network model training and human face comparison method, and terminal and medium Download PDFInfo
- Publication number
- WO2019192121A1 WO2019192121A1 PCT/CN2018/100152 CN2018100152W WO2019192121A1 WO 2019192121 A1 WO2019192121 A1 WO 2019192121A1 CN 2018100152 W CN2018100152 W CN 2018100152W WO 2019192121 A1 WO2019192121 A1 WO 2019192121A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- face
- feature
- face image
- neural network
- normalized
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Definitions
- the present application relates to the field of image recognition technologies, and in particular, to a two-channel neural network model training and a face comparison method, a terminal, and a medium.
- biometrics have been widely used in identity authentication because of their portability, loss, no forgetting, no borrowing, and no misappropriation.
- face recognition technology is the most direct, friendly and convenient, and is our ideal choice.
- Face matching is a sub-area of face recognition. Face comparison is to judge whether two faces are the same person. The most common scene is to judge whether the certificate is the person or not. Face recognition is given a face. The picture, and then judge who this person is, its essence is equivalent to multiple face comparisons.
- the face image in the dynamic environment has many effects such as insufficient illumination, occlusion, insufficient resolution, and incorrect posture, the face matching in the dynamic environment is very difficult, resulting in a decrease in the accuracy of face comparison.
- a first aspect of the present application provides a two-channel neural network model training method, the method comprising:
- each original face image has a size of 182*182;
- the dual channel neural network model training End and update the weights and offsets in the two-channel neural network model.
- a second aspect of the present application provides a method for face alignment using the two-channel neural network model, the method comprising:
- the histogram of the target user After receiving the face image of the target user to be compared, the histogram of the target user is subjected to histogram equalization processing to obtain a histogram equalized face image of the target user;
- the face comparison result is determined according to the relationship between the calculated similarity and the preset confidence threshold.
- a third aspect of the present application provides a terminal, the terminal comprising a processor, the processor is configured to implement the dual channel neural network model training method or implement the face ratio when executing computer readable instructions stored in a memory The method.
- a fourth aspect of the present application provides a non-volatile readable storage medium having stored thereon computer readable instructions, the computer readable instructions being implemented by a processor to implement the The two-channel neural network model training method or the face matching method is implemented.
- the two-channel neural network model training method, the face comparison method, the terminal and the medium described in the present application apply the trained two-channel neural network model to the face comparison, and can solve the existence of the face image in the dynamic environment. Insufficient illumination, occlusion, insufficient resolution, incorrect posture, etc.; only a small number of samples are needed in the training phase of the two-channel neural network model, thus solving the problem of the data volume requirement of the algorithm and increasing the practicability of the algorithm. Improve the accuracy and efficiency of witness verification.
- FIG. 1 is a flowchart of a two-channel neural network model training method according to Embodiment 1 of the present application.
- Embodiment 2 is a flow chart of a two-channel neural network model training method provided by Embodiment 2 of the present application.
- FIG. 3 is a flowchart of a face matching method according to Embodiment 3 of the present application.
- Embodiment 4 is a flowchart of a face matching method provided in Embodiment 4 of the present application.
- FIG. 5 is a structural diagram of a two-channel neural network model training apparatus according to Embodiment 5 of the present application.
- FIG. 6 is a structural diagram of a face matching device according to Embodiment 6 of the present application.
- FIG. 7 is a schematic diagram of a terminal provided in Embodiment 7 of the present application.
- the face matching method of the present application is applied to one or more terminals.
- the two-channel neural network model training method and the face comparison method can also be applied to a hardware environment composed of a terminal and a server connected to the terminal through a network.
- the two-channel neural network model training method and the face comparison method of the embodiment of the present application may be performed by the server at the same time, or may be performed by the terminal at the same time; or may be performed by the server and the terminal together, for example, the dual channel nerve
- the network model training method is performed by a server, and the face matching method is executed by a terminal, or the face matching method is executed by a server, and the two-channel neural network model training method is executed by a terminal.
- This application is not limited herein.
- FIG. 1 is a flowchart of a two-channel neural network model training method according to Embodiment 1 of the present application.
- the obtaining of the original face image may include the following two methods:
- the resolution of the photographing device for example, camera, camera, etc.
- the face of different people can be photographed in a dynamic environment to obtain the original face image of size 182*182;
- the LFW data set was created to investigate the face recognition problem in an unrestricted environment, which contains more than 13,000 face images, all of which are from the Internet, not the lab environment.
- each original face image is randomly cut into five first face images of the same size, and each first face image is Both contain face areas.
- Randomly crop each original face image to obtain the first face image under arbitrary angle and arbitrary illumination conditions is the process of collecting the original face image in the simulated dynamic environment; secondly, by cutting In this way, multiple first face images can be obtained, thereby increasing the number of face sample data sets, and more face sample data sets can improve the robustness of the two-channel neural network model.
- the specific steps of performing histogram equalization processing on the first face image in the face sample data set include:
- n is the sum of the number of pixels in the first face picture
- n k is the number of pixels of the gray level r k
- L is the total number of gray levels in the first face picture
- the calculated pixel grayscale distribution density is transformed according to formula (1-2).
- the gray value is mapped to obtain a new pixel gray level, and the new pixel gray level is rounded to obtain a transformed new pixel gray level.
- Performing a histogram equalization process on each of the first face images in the face sample data set which can further enhance the contrast of the first face image in the face sample data set, especially for exposure
- the first face image that is excessive or underexposed can better maintain the detail information of the face region in the first face image after the histogram equalization processing.
- the color picture needs to be The color space is converted into an HSV (Hue Chroma, Value Saturation) color space from the RGB color space, and then the histogram equalization process is performed on the V component.
- HSV Human Chroma, Value Saturation
- the first deep neural network and the second deep neural network are neural networks having the same number of layers (for example, 200 layers).
- the first depth neural network and the second deep neural network adopt a perceptron-based Inception-Resnet-V2 (Inception-V2 and Resnet combined neural network) neural network.
- the perceptron-based Inception-Resnet-V2 neural network is a prior art, and the present application does not repeat here.
- the step 104 may further include: the dimension of the first facial feature extracted by the first depth neural network is 64 dimensions, and the second person extracted by the second deep neural network The dimension of the face feature is 64 dimensions.
- the first facial feature and the second facial feature may be normalized by a normalization function, and the normalization function may be a Euclidean distance function or a Manhattan Distance function, minimum absolute error.
- the first normalized face feature is obtained by using the Euclidean distance function on the first facial feature, and the second facial feature is normalized to obtain a second normalized facial feature.
- Normalization can compress light, shadows, and edges so that the first normalized face feature and the second normalized face feature are robust to illumination, shadow, and edge changes, further enhancing the two-channel neural network. Model robustness.
- normalization using the Euclidean distance function can avoid over-fitting of the two-channel neural network model, thereby improving the generalization ability of the two-channel neural network model, and optimizing the weight and offset of the subsequent two-channel neural network model. Become more stable and fast.
- the preset connection rule may be a cross connection.
- the first normalized face feature is (x 1 , x 2 , x 3 , . . . , x m )
- the second normalized face feature is (y 1 , y 2 , y 3 , . . . , y m )
- the final face features obtained by the cross-connection connection are expressed as (x 1 , y 1 , x 2 , y 2 , x 3 , y 3 , . . . , x m , y m ).
- the preset connection rules may be sequential connections.
- the first normalized face feature is (x 1 , x 2 , x 3 , . . . , x m )
- the second normalized face feature is (y 1 , y 2 , y 3 , . . . , y m )
- the final face features obtained by sequential connection are expressed as (x 1 , x 2 , x 3 , . . . , x m , y 1 , y 2 , y 3 , . . . , y m ).
- 107 normalize the final facial feature representation and input to a preset loss function to calculate a loss function value.
- the loss function value is less than or equal to a preset loss function threshold, the dual channel neural network model training Ending and updating the weights and offsets in the two-channel neural network model; when the loss function value is greater than a preset loss function threshold, acquiring more original face images and re-executing based on the increased original face images.
- the preset loss function L is represented by the following formula (1-3):
- L S represents the result of the cross-entropy loss function (softmax loss). Represents the actual output of the yi person, L C represents the result of the center loss function, cyi represents the feature center of the yi person, and xi represents the feature before the fully connected layer.
- the preset loss function can make the distance between different face images larger, and the distance between different face images of the same person is reduced, thereby improving the classification of the two-channel neural network model.
- the dual-channel neural network model trained in the present application can significantly enhance the feature extraction capability of the face image in the input face sample data set and the face image processed by the histogram equalization process. It can reduce the risk of performance degradation when the network level is deepened, and improve the robustness of the two-channel neural network model, which is convenient for improving the recognition rate of face recognition or face comparison.
- Embodiment 2 is a flow chart of a two-channel neural network model training method provided by Embodiment 2 of the present application.
- the step 201 in this embodiment is the same as the step 101 in the first embodiment, and the application is not described in detail herein.
- the second face image size obtained after random cropping is 171*160, which not only can maintain the facial detail information in the original face image to the greatest extent, but also utilizes the original face image.
- the training time of the two-channel neural network model can be shortened, and the training speed of the two-channel neural network model can be improved.
- the pre-stored face detection algorithm may be one of the following algorithms: a template-based face detection method, an artificial neural network-based face detection method, a model-based face detection method, and a skin color-based face detection method. The method or the face detection method based on the feature sub-face, and the like.
- the pre-stored face detection algorithm is not limited to the above enumeration, and any algorithm capable of detecting whether the face image is included in the cropped face image may be referred to herein.
- the face detection algorithm pre-stored in this embodiment is a prior art, and will not be described in detail herein.
- each of the second face images in the cropped second face image includes a face region
- performing the following step 204 when determining each of the cropped second face images or When the face area is not included in the partial face picture, the following step 205 is performed.
- the face region that is not included in the cropped second face image includes the following two situations: there is no face region at all; only a part of the face region exists.
- the step 204 in this embodiment is the same as the step 103 in the first embodiment, and the application is not described in detail herein.
- the step 206 in this embodiment is the same as the step 104 in the first embodiment, and the application is not described in detail herein.
- the step 207 in this embodiment is the same as the step 105 in the first embodiment, and the application is not described in detail herein.
- the step 208 in this embodiment is the same as the step 106 in the first embodiment, and the application is not described in detail herein.
- the step 209 in this embodiment is the same as the step 107 in the first embodiment, and the present application will not be described in detail herein.
- the two-channel neural network model trained in the present application detects each of the cropped second face images by a pre-stored face detection algorithm to ensure that each second face image is included. All the face regions can ensure the validity of the extracted face features, which can further improve the robustness of the two-channel neural network model, and facilitate the subsequent recognition of face recognition or face comparison.
- FIG. 3 is a flowchart of a face matching method according to Embodiment 3 of the present application.
- the feature vector A represents the final feature representation of the target user
- the feature vector B represents the final feature representation of the registered user.
- the cosine value between the feature vector A and each of the feature vectors B in the database is calculated according to the equation (2-1), and the range of the calculated cosine value is [-1, 1].
- the SIM calculated according to equation (2-2) is the similarity, and the value of the similarity is normalized to [0, 1].
- the results of the face alignment include alignment results and comparison failure results.
- the calculated similarity is greater than or equal to the preset confidence threshold, it is determined that the user matches, that is, the registered user who is the same person as the target user is matched in the database.
- the calculated similarity is less than the preset confidence threshold, it is determined that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
- the preset confidence threshold is 0.5.
- the dual-channel neural network model is used for face comparison, which can improve the accuracy of face comparison, shorten the time of face comparison, and improve the efficiency of face comparison.
- Embodiment 4 is a flowchart of a face matching method provided in Embodiment 4 of the present application.
- the pre-processing may include, but is not limited to, image denoising, illumination normalization, pose calibration, grayscale normalization, and the like.
- the face image of the target user After receiving the face image of the target user to be compared, the face image of the target user is pre-processed to further improve the face image of the target user, so as to extract more discriminative power. Face features.
- 403 Input the pre-processed target user's face image into the first deep neural network in the trained two-channel neural network model to extract the third facial feature, and simultaneously equalize the target user's histogram.
- the fourth face feature is extracted from the second deep neural network in the trained two-channel neural network model.
- the pre-set sorting method includes, but is not limited to, sorting from large to small, and sorting from small to large.
- the results of the face alignment include alignment results and comparison failure results.
- the maximum similarity is greater than or equal to the preset confidence threshold, it is determined that the user matches, that is, the registered user who is the same person as the target user is matched in the database.
- the maximum similarity is less than the preset confidence threshold, it is determined that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
- the preset confidence threshold is 0.5.
- the user comparison is determined. by.
- the maximum similarity is less than the preset confidence threshold, it is determined that the user comparison fails. This can shorten the comparison time, and only need to compare once to determine whether the target user is a registered user.
- the face matching method described in the present application can extract a better facial feature representation by using a two-channel neural network model, and a better face can be obtained by taking a shorter face comparison time. Compare effects.
- FIG. 5 is a functional block diagram of a preferred embodiment of the dual channel neural network model training device of the present application.
- the two-channel neural network model training device 50 operates in the terminal 7.
- the dual channel neural network model training device 50 can include a plurality of functional modules comprised of computer readable instruction code segments. Each of the computer readable instruction code segments in the dual channel neural network model training device 50 may be stored in the memory 71 and executed by the at least one processor 72 for execution (see Figures 1 and 2 for details). Description) Training for a two-channel neural network model.
- the dual channel neural network model training device 50 can be divided into multiple functional modules according to the functions performed by the dual channel neural network model.
- the function module may include: an obtaining module 501, a cropping module 502, a first histogram processing module 503, a first feature extraction module 504, a first normalization module 505, a first feature connection module 506, and a model update module 507. And a face detection module 508.
- the obtaining module 501 is configured to obtain a raw face image of different people, and the size of each original face image is 182*182.
- the cropping module 502 is configured to randomly cut each original face image into a preset number of first face images to obtain a face sample data set.
- the preset number can be adjusted or modified according to actual needs.
- the cropping module 502 randomly cuts each original face image into five first face images of the same size, each of which The first face image contains a face area.
- Randomly crop each original face image to obtain the first face image under arbitrary angle and arbitrary illumination conditions is the process of collecting the original face image in the simulated dynamic environment; secondly, by cutting In this way, multiple first face images can be obtained, thereby increasing the number of face sample data sets, and more face sample data sets can improve the robustness of the two-channel neural network model.
- the histogram processing module 503 is configured to perform histogram equalization processing on each first face image in the face sample data set to obtain a histogram equalized face image.
- the first feature extraction module 504 is configured to input each first face image in the face sample data into the first depth neural network to extract the first facial feature, and simultaneously equalize the corresponding histogram.
- the picture input second depth neural network extracts the second face feature.
- the first deep neural network and the second deep neural network are neural networks having the same number of layers (for example, 200 layers).
- the first depth neural network and the second deep neural network adopt a perceptron-based Inception-Resnet-V2 (Inception-V2 and Resnet combined neural network) neural network.
- the perceptron-based Inception-Resnet-V2 neural network is a prior art, and the present application does not repeat here.
- the first feature extraction module 504 is further configured to: the dimension of the first facial feature extracted by the first depth neural network is 64 dimensions, and the second depth neural network extracts The dimension of the second face feature is 64 dimensions.
- a first normalization module 505 configured to normalize the first facial feature to obtain a first normalized facial feature, and normalize the second facial feature to obtain a second normalized Face features.
- a first feature connection module 506 configured to connect the first normalized face feature and the second normalized face feature according to a preset connection rule to obtain a new normalized face feature, as a final face Feature representation.
- the model update module 507 is configured to normalize the final facial feature representation and input the value to the preset loss function to calculate a loss function value.
- the loss function value is less than or equal to a preset loss function threshold
- the double The channel neural network model training ends and updates the weights and offsets in the two-channel neural network model; when the loss function value is greater than the preset loss function threshold, more original face images are acquired, and based on the increased original The face picture re-executes the above modules 501-507.
- the cropping module 502 is further configured to randomly cut each original face image into a preset number of second face images to obtain a face sample data set, where each of the face sample data sets The size of the second face image is 171*160.
- the second face image size obtained after random cropping is 171*160, which not only can maintain the facial detail information in the original face image to the greatest extent, but also utilizes the original face image.
- the training time of the two-channel neural network model can be shortened, and the training speed of the two-channel neural network model can be improved.
- the dual-channel neural network model training device 50 further includes: a face detection module 508, configured to detect each of the cropped second face images by using a pre-stored face detection algorithm, and determine the cropping Whether or not the face area is included in each second face picture in all the second face pictures.
- a face detection module 508 configured to detect each of the cropped second face images by using a pre-stored face detection algorithm, and determine the cropping Whether or not the face area is included in each second face picture in all the second face pictures.
- the pre-stored face detection algorithm may be one of the following algorithms: a template-based face detection method, an artificial neural network-based face detection method, a model-based face detection method, and a skin color-based face detection method. The method or the face detection method based on the feature sub-face, and the like.
- the pre-stored face detection algorithm is not limited to the above enumeration, and any algorithm capable of detecting whether the face image is included in the cropped face image may be referred to herein.
- the face detection algorithm pre-stored in this embodiment is a prior art, and will not be described in detail herein.
- the first histogram processing module 503 is further configured to: when the face detection module 508 determines that each of the second face images in the cropped second face image includes a face region, Histogram equalization processing is performed on each second face image in the face sample data set to obtain a histogram equalization face image.
- the cropping module 502 is further configured to: when the face detection module 508 determines that each of the second face images that are cropped includes no face region in the face image, the pair does not include The original face image corresponding to the second face image of the face region is randomly cropped until the preset number of second face images are obtained.
- the face region that is not included in the cropped second face image includes the following two situations: there is no face region at all; only a part of the face region exists.
- the first normalization module 505 is further configured to input each second face image in the face sample data into the first depth neural network to extract the first facial feature, and correspondingly The histogram equalization face image input second depth neural network extracts the second face feature.
- the dual-channel neural network model trained in the present application can significantly enhance the feature extraction capability of the face image in the input face sample data set and the face image processed by the histogram equalization process. It mitigates the risk of performance degradation when the network level is deepened, and improves the robustness of the two-channel neural network model.
- each second face image that is cropped is detected by a pre-stored face detection algorithm to ensure each piece.
- the second face image contains all the face regions, which can ensure the validity of the extracted face features, and can further improve the robustness of the two-channel neural network model, and facilitate the subsequent improvement of face recognition or face comparison. Recognition rate.
- FIG. 6 is a functional block diagram of a preferred embodiment of the applicant's face matching device.
- the face matching device 60 operates in the terminal 7.
- the face matching device 60 can include a plurality of functional modules comprised of computer readable instruction code segments.
- the instruction codes of the respective computer readable instruction segments in the face matching device 60 may be stored in the memory 71 and executed by the at least one processor 72 for execution (see Figures 3 and 4 for details). Description) The comparison of faces.
- the face matching device 60 of the terminal 7 can be divided into a plurality of functional modules according to the functions performed by the terminal 7.
- the function module may include: a second histogram processing module 601, a second feature extraction module 602, a second normalization module 603, a second feature connection module 604, a calculation module 605, a result determination module 606, and a pre-processing module 607. And sorting module 608.
- the second histogram processing module 601 is configured to perform histogram equalization processing on the face image of the target user after receiving the face image of the target user to be compared by the face, and obtain a histogram equalization of the target user. Face picture.
- a second feature extraction module 602 configured to input a face image of the target user into a first depth neural network in the trained two-channel neural network model, and extract a third facial feature, and simultaneously The fourth facial feature is extracted from the second deep neural network in the two-channel neural network model trained by the histogram equalization face image input;
- the second normalization module 603 is configured to normalize the third facial feature to obtain a third normalized facial feature, and normalize the fourth facial feature to obtain a fourth normalized facial feature. ;
- the second feature connection module 604 is configured to connect the third normalized face feature and the fourth normalized face feature according to the preset connection rule to obtain a new normalized face feature. The final feature representation of the target user.
- the calculating module 605 is configured to calculate a similarity between the final feature representation of the target user and the final feature representation of the registered user in the database.
- the result determining module 606 is configured to determine a face comparison result according to a relationship between the calculated similarity and a preset confidence threshold.
- the results of the face alignment include alignment results and comparison failure results.
- the result determination module 606 determines that the user matches, that is, the registered user who is the same person as the target user is matched in the database.
- the result determination module 606 determines that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
- the preset confidence threshold is 0.5.
- the face matching device 60 further includes: a pre-processing module 607, configured to pre-process the face image of the target user after receiving the face image of the target user to be compared by the face deal with.
- the pre-processing may include, but is not limited to, image denoising, illumination normalization, pose calibration, grayscale normalization, and the like.
- the face image of the target user After receiving the face image of the target user to be compared, the face image of the target user is pre-processed to further improve the face image of the target user, so as to extract more discriminative power. Face features.
- the second histogram processing module 601 further performs histogram equalization processing on the pre-processed target user's face image to obtain a histogram equalized face image of the target user.
- the second feature extraction module 602 is further configured to input the face image of the pre-processed target user into the first deep neural network in the trained two-channel neural network model to extract the third facial feature. And extracting the fourth facial feature from the second deep neural network in the trained two-channel neural network model by inputting the histogram equalized face image of the target user;
- the face matching device 60 further includes: a sorting module 608, configured to sort the calculated similarities according to a preset sorting method.
- the pre-set sorting method includes, but is not limited to, sorting from large to small, and sorting from small to large.
- the result determining module 606 is further configured to compare a relationship between the maximum similarity and a preset confidence threshold to determine a face comparison result.
- the results of the face alignment include alignment results and comparison failure results.
- the result determination module 606 determines that the user has passed the comparison, ie, the registered user who is the same person as the target user is matched in the database.
- the result determination module 606 determines that the user comparison failed, ie, the registered user who is the same person as the target user is not matched in the database.
- the face matching method described in the present application can extract a better facial feature representation by using a two-channel neural network model, and a better face can be obtained by taking a shorter face comparison time. Compare effects.
- FIG. 7 it is a hardware structure diagram of a terminal for implementing the dual channel neural network model training method and/or the face matching method described in the present application.
- the terminal 7 includes a memory 71, at least one processor 72, and at least one communication bus 73.
- the structure of the terminal 7 shown in FIG. 7 does not constitute a limitation of the embodiment of the present application, and may be a bus type structure or a star structure, and the terminal 7 may further include a ratio More or less other hardware or software, or different component arrangements.
- the memory 71 is configured to store computer readable instruction code and various data, such as the dual channel neural network model training device 50 and the face matching device 60 installed in the terminal 7, and The high speed, automatic completion of computer readable instructions or data access during the operation of the terminal 7.
- the at least one processor 72 may be composed of an integrated circuit, for example, may be composed of a single packaged integrated circuit, or may be composed of multiple integrated circuits of the same function or different functions, including one. Or a combination of a plurality of central processing units, microprocessors, digital processing chips, graphics processors, and various control chips.
- the at least one processor 72 is a control core of the terminal 7, connecting various components of the entire terminal 7 using various interfaces and lines, by running or executing computer readable instructions or modules stored in the memory 71, and The data stored in the memory 71 is invoked to perform various functions and processing data of the terminal 7, such as executing the two-channel neural network model training device 50 and/or the face matching device 60.
- the at least one communication bus 73 is arranged to effect connection communication between the memory 71, the at least one processor 72, and the like.
- the functional modules in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
- the above integrated unit can be implemented in the form of hardware or in the form of hardware plus software function modules.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Image Analysis (AREA)
Abstract
Description
本申请要求于2018年04月04日提交中国专利局,申请号为201810299180.6发明名称为“双通道神经网络模型训练及人脸比对方法、终端及介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims to be filed on April 4, 2018, the Chinese Patent Office, the application number is 201101829180.6. The invention is entitled "Double-channel neural network model training and face comparison method, terminal and medium" priority of Chinese patent application, all of which The content is incorporated herein by reference.
本申请涉及图像识别技术领域,具体涉及一种双通道神经网络模型训练及人脸比对方法、终端及介质。The present application relates to the field of image recognition technologies, and in particular, to a two-channel neural network model training and a face comparison method, a terminal, and a medium.
最近几年,随着机器学习和深度学习的发展,生物特征凭借便于携带、不会丢失、不会遗忘、不会借用和不会盗用等特点,被广泛应用在身份认证中。人脸识别技术作为生物特征识别技术之一,它具有最直接、友好和方便的特点,是我们的理想选择。人脸比对是人脸识别的子领域,人脸比对是判断两张人脸图片是不是同一个人,最常用的场景是判断证件是不是本人,人脸识别则是给定一张人脸图片,然后判断这个人是谁,其实质相当于多次的人脸比对。In recent years, with the development of machine learning and deep learning, biometrics have been widely used in identity authentication because of their portability, loss, no forgetting, no borrowing, and no misappropriation. As one of the biometric recognition technologies, face recognition technology is the most direct, friendly and convenient, and is our ideal choice. Face matching is a sub-area of face recognition. Face comparison is to judge whether two faces are the same person. The most common scene is to judge whether the certificate is the person or not. Face recognition is given a face. The picture, and then judge who this person is, its essence is equivalent to multiple face comparisons.
由于动态环境下的人脸图片存在光照不足、遮挡、分辨率不够、姿态不正确等多种影响,使得动态环境下的人脸比对难度非常大,造成人脸比对准确度下降。Because the face image in the dynamic environment has many effects such as insufficient illumination, occlusion, insufficient resolution, and incorrect posture, the face matching in the dynamic environment is very difficult, resulting in a decrease in the accuracy of face comparison.
发明内容Summary of the invention
鉴于以上内容,有必要提出一种双通道神经网络模型训练方法及人脸比对方法、终端及介质,其可以训练出适合人脸比对的分类模型,训练出的特征的区分能力较强,获得较佳的人脸比对效果。In view of the above, it is necessary to propose a two-channel neural network model training method, a face matching method, a terminal and a medium, which can train a classification model suitable for face matching, and the trained features have strong distinguishing ability. Get better face matching effect.
本申请的第一方面提供一种双通道神经网络模型训练方法,所述方法包括:A first aspect of the present application provides a two-channel neural network model training method, the method comprising:
a.获取不同人的一张原始人脸图片,每张原始人脸图片的尺寸为182*182;a. obtaining a primitive face image of different people, each original face image has a size of 182*182;
b.将每张原始人脸图片进行随机裁剪成预设数量的人脸图片,得到人脸样本数据集;b. randomly cutting each original face image into a preset number of face images to obtain a face sample data set;
c.分别对所述人脸样本数据集中的每张人脸图片进行直方图均衡化处理,得到直方图均衡化人脸图片;c. performing histogram equalization processing on each face image in the face sample data set to obtain a histogram equalization face image;
d.将人脸样本数据中的每张所述人脸图片输入第一深度神经网络中提取出第一人脸特征,同时将对应的直方图均衡化人脸图片输入第二深度神经网络中提取出第二人脸特征;d. inputting each face image in the face sample data into the first depth neural network to extract the first face feature, and inputting the corresponding histogram equalized face image into the second depth neural network for extracting a second face feature;
e.对所述第一人脸特征进行归一化得到第一归一化人脸特征,同时对所述第二人脸特征进行归一化得到第二归一化人脸特征;e. normalizing the first facial feature to obtain a first normalized facial feature, and normalizing the second facial feature to obtain a second normalized facial feature;
f.根据预设连接规则连接所述第一归一化人脸特征和所述第二归一化人脸特征得到新的归一化人脸特征,作为最终人脸特征表示;及f. connecting the first normalized face feature and the second normalized face feature according to a preset connection rule to obtain a new normalized face feature as a final face feature representation;
g.将所述最终人脸特征表示进行归一化后输入至预先设置的损失函数中计算损失函数值,当损失函数值小于或等于预先设置的损失函数阈值时,则 双通道神经网络模型训练结束并更新所述双通道神经网络模型中的权重和偏置。g. normalizing the final facial feature representation and inputting it into a preset loss function to calculate a loss function value. When the loss function value is less than or equal to a preset loss function threshold, the dual channel neural network model training End and update the weights and offsets in the two-channel neural network model.
本申请的第二方面提供一种利用所述的双通道神经网络模型进行人脸比对的方法,所述方法包括:A second aspect of the present application provides a method for face alignment using the two-channel neural network model, the method comprising:
在收到待进行人脸比对的目标用户的人脸图片后,对目标用户的人脸图片进行直方图均衡化处理,得到目标用户的直方图均衡化人脸图片;After receiving the face image of the target user to be compared, the histogram of the target user is subjected to histogram equalization processing to obtain a histogram equalized face image of the target user;
将所述目标用户的人脸图片输入训练好的双通道神经网络模型中的第一深度神经网络中提取出第三人脸特征,同时将所述目标用户的直方图均衡化人脸图片输入训练好的双通道神经网络模型中的第二深度神经网络中提取出第四人脸特征;Importing a face image of the target user into a first deep neural network in the trained two-channel neural network model to extract a third facial feature, and simultaneously mapping the target user's histogram equalized face image into the training The fourth facial feature is extracted from the second deep neural network in the good two-channel neural network model;
对所述第三人脸特征进行归一化得到第三归一化人脸特征,同时对所述第四人脸特征进行归一化得到第四归一化人脸特征;Normalizing the third facial feature to obtain a third normalized facial feature, and normalizing the fourth facial feature to obtain a fourth normalized facial feature;
根据所述预设连接规则连接所述第三归一化人脸特征和所述第四归一化人脸特征得到新的归一化人脸特征,作为所述目标用户的最终特征表示;And connecting the third normalized face feature and the fourth normalized face feature according to the preset connection rule to obtain a new normalized face feature as a final feature representation of the target user;
计算所述目标用户的最终特征表示与数据库中注册用户的最终特征表示之间的相似度;及Calculating a similarity between a final feature representation of the target user and a final feature representation of a registered user in the database; and
根据计算得到的相似度与预先设置的置信度阈值之间的关系,确定人脸比对结果。The face comparison result is determined according to the relationship between the calculated similarity and the preset confidence threshold.
本申请的第三方面提供一种终端,所述终端包括处理器,所述处理器用于执行存储器中存储的计算机可读指令时实现所述双通道神经网络模型训练方法或者实现所述人脸比对方法。A third aspect of the present application provides a terminal, the terminal comprising a processor, the processor is configured to implement the dual channel neural network model training method or implement the face ratio when executing computer readable instructions stored in a memory The method.
本申请的第四方面提供一种非易失性可读存储介质,所述非易失性可读存储介质上存储有计算机可读指令,所述计算机可读指令被处理器执行时实现所述双通道神经网络模型训练方法或者实现所述人脸比对方法。A fourth aspect of the present application provides a non-volatile readable storage medium having stored thereon computer readable instructions, the computer readable instructions being implemented by a processor to implement the The two-channel neural network model training method or the face matching method is implemented.
本申请所述的双通道神经网络模型训练方法及人脸比对方法、终端及介质,将训练出的双通道神经网络模型应用在人脸比对上,能够解决动态环境下的人脸图片存在光照不足、遮挡、分辨率不够、姿态不正确等多种影响;在双通道神经网络模型训练阶段只需要少量样本即可,从而解决了算法对数据量的需求问题,增加了算法的实用性,提高了人证验证的准确率及效率。The two-channel neural network model training method, the face comparison method, the terminal and the medium described in the present application apply the trained two-channel neural network model to the face comparison, and can solve the existence of the face image in the dynamic environment. Insufficient illumination, occlusion, insufficient resolution, incorrect posture, etc.; only a small number of samples are needed in the training phase of the two-channel neural network model, thus solving the problem of the data volume requirement of the algorithm and increasing the practicability of the algorithm. Improve the accuracy and efficiency of witness verification.
图1是本申请实施例一提供的双通道神经网络模型训练方法的流程图。FIG. 1 is a flowchart of a two-channel neural network model training method according to Embodiment 1 of the present application.
图2是本申请实施例二提供的双通道神经网络模型训练方法的流程图。2 is a flow chart of a two-channel neural network model training method provided by Embodiment 2 of the present application.
图3是本申请实施例三提供的人脸比对方法的流程图。FIG. 3 is a flowchart of a face matching method according to Embodiment 3 of the present application.
图4是本申请实施例四提供的人脸比对方法的流程图。4 is a flowchart of a face matching method provided in Embodiment 4 of the present application.
图5是本申请实施例五提供的双通道神经网络模型训练装置的结构图。FIG. 5 is a structural diagram of a two-channel neural network model training apparatus according to Embodiment 5 of the present application.
图6是本申请实施例六提供的人脸比对装置的结构图。FIG. 6 is a structural diagram of a face matching device according to Embodiment 6 of the present application.
图7是本申请实施例七提供的终端的示意图。FIG. 7 is a schematic diagram of a terminal provided in
本申请的人脸比对方法应用在一个或者多个终端中。所述双通道神经网络模型训练方法及人脸比对方法也可以应用于由终端和通过网络与所述终端 进行连接的服务器所构成的硬件环境中。本申请实施例的双通道神经网络模型训练方法及人脸比对方法可以同时由服务器来执行,也可以同时由终端来执行;还可以是由服务器和终端共同执行,比如,所述双通道神经网络模型训练方法由服务器来执行,所述人脸比对方法由终端来执行,或者,所述人脸比对方法由服务器来执行,所述双通道神经网络模型训练方法由终端来执行。本申请在此不加以限制。The face matching method of the present application is applied to one or more terminals. The two-channel neural network model training method and the face comparison method can also be applied to a hardware environment composed of a terminal and a server connected to the terminal through a network. The two-channel neural network model training method and the face comparison method of the embodiment of the present application may be performed by the server at the same time, or may be performed by the terminal at the same time; or may be performed by the server and the terminal together, for example, the dual channel nerve The network model training method is performed by a server, and the face matching method is executed by a terminal, or the face matching method is executed by a server, and the two-channel neural network model training method is executed by a terminal. This application is not limited herein.
实施例一Embodiment 1
图1是本申请实施例一提供的双通道神经网络模型训练方法的流程图。FIG. 1 is a flowchart of a two-channel neural network model training method according to Embodiment 1 of the present application.
101:获取不同人的一张原始人脸图片,每张原始人脸图片的尺寸为182*182。101: Obtain a raw face picture of different people, and the size of each original face picture is 182*182.
本较佳实施例中,所述原始人脸图片的获取可包括以下两种方法:In the preferred embodiment, the obtaining of the original face image may include the following two methods:
(1)事先设置拍照设备(例如,摄像机、照相机等)的分辨率为182*182,则在动态环境下对不同人的人脸进行拍摄即可得到尺寸为182*182的原始人脸图片;(1) If the resolution of the photographing device (for example, camera, camera, etc.) is set to 182*182 in advance, the face of different people can be photographed in a dynamic environment to obtain the original face image of size 182*182;
(2)从人脸数据集中获取不同人的原始人脸图片,将原始人脸图片设置为182*182的尺寸。(2) Obtain the original face image of different people from the face data set, and set the original face picture to the size of 182*182.
所述LFW数据集是为了研究非限制环境下的人脸识别问题而建立,其中包含了超过13000张人脸图像,人脸图像全部来自于Internet,而不是实验室环境。The LFW data set was created to investigate the face recognition problem in an unrestricted environment, which contains more than 13,000 face images, all of which are from the Internet, not the lab environment.
将原始人脸图片分辨率设置为182*182,不仅能获得较多的人脸面部细节信息,且能适配现有通用的传输、存储和管理系统及设备。Setting the original face image resolution to 182*182 not only provides more facial detail information, but also adapts to existing general transmission, storage, and management systems and devices.
102:将每张原始人脸图片进行随机裁剪成预设数量的第一人脸图片,得到人脸样本数据集。102: Randomly crop each original face image into a preset number of first face images to obtain a face sample data set.
所述预设数量可以根据实际需要自行调整或修改,本较佳实施例中,将每张原始人脸图片进行随机裁剪成大小相同的5张第一人脸图片,每张第一人脸图片均包含人脸区域。The preset number can be adjusted or modified according to actual needs. In the preferred embodiment, each original face image is randomly cut into five first face images of the same size, and each first face image is Both contain face areas.
将每张原始人脸图片进行随机裁剪,可得到任意角度、任意光照等条件下的第一人脸图片,随机的过程即是模拟动态环境下采集原始人脸图片的过程;其次,通过裁剪的方式可得到多张第一人脸图片,从而增加人脸样本数据集的数量,较多的人脸样本数据集能够提高双通道神经网络模型的鲁棒性。Randomly crop each original face image to obtain the first face image under arbitrary angle and arbitrary illumination conditions. The random process is the process of collecting the original face image in the simulated dynamic environment; secondly, by cutting In this way, multiple first face images can be obtained, thereby increasing the number of face sample data sets, and more face sample data sets can improve the robustness of the two-channel neural network model.
103:分别对所述人脸样本数据集中的每张第一人脸图片进行直方图均衡化处理,得到直方图均衡化人脸图片。103: Perform a histogram equalization process on each first face image in the face sample data set to obtain a histogram equalization face image.
本实施例中,对所述人脸样本数据集中的第一人脸图片进行直方图均衡化处理的具体步骤包括:In this embodiment, the specific steps of performing histogram equalization processing on the first face image in the face sample data set include:
1)对所述第一人脸图片进行像素灰度统计;1) performing pixel grayscale statistics on the first face image;
统计所述第一人脸图片中每个像素灰度级出现的次数,将相同的灰度级像素点数量进行累加。Counting the number of occurrences of each pixel gray level in the first face image, and accumulating the same number of gray level pixels.
2)计算像素灰度分布密度;2) Calculate the pixel gray distribution density;
根据公式(1-1)计算第k灰度级在整张所述第一人脸图片中出现的概率Pr(r k), Calculating the probability Pr(r k ) of the kth gray level appearing in the entire first face picture according to the formula (1-1),
其中,n是所述第一人脸图片中像素个数的总和,n k是灰度级r k的像素个数,L是所述第一人脸图片中灰度级的总数。 Where n is the sum of the number of pixels in the first face picture, n k is the number of pixels of the gray level r k , and L is the total number of gray levels in the first face picture.
3)根据累积分布函数对所计算的像素灰度分布密度进行变换;3) transforming the calculated grayscale distribution density of the pixel according to the cumulative distribution function;
根据公式(1-2)对所计算的像素灰度分布密度进行变换。The calculated pixel grayscale distribution density is transformed according to formula (1-2).
4)计算变换后的新像素灰度;4) calculating the converted new pixel gray scale;
对灰度值进行映射得到新像素灰度,并对所述新像素灰度进行取整,得到变换后的新像素灰度。The gray value is mapped to obtain a new pixel gray level, and the new pixel gray level is rounded to obtain a transformed new pixel gray level.
对所述人脸样本数据集中的每张所述第一人脸图片进行直方图均衡化处理,能够进一步增强所述人脸样本数据集中的所述第一人脸图片的对比度,尤其是对于曝光过度或者曝光不足的所述第一人脸图片,经过所述直方图均衡化处理后能够更好的保持所述第一人脸图片中的人脸区域的细节信息。Performing a histogram equalization process on each of the first face images in the face sample data set, which can further enhance the contrast of the first face image in the face sample data set, especially for exposure The first face image that is excessive or underexposed can better maintain the detail information of the face region in the first face image after the histogram equalization processing.
优选地,若人脸样本数据集中的所述第一人脸图片为彩色图片,由于无法直接对彩色图片中的R、G、B三个分量进行直方图均衡化处理,因而需要将彩色图片的颜色空间由RGB色彩空间转换成HSV(Hue色度,Saturation饱和度,Value纯度)色彩空间,然后对V分量进行所述直方图均衡化处理。Preferably, if the first face image in the face sample data set is a color picture, since the histogram equalization processing of the three components R, G, and B in the color picture cannot be directly performed, the color picture needs to be The color space is converted into an HSV (Hue Chroma, Value Saturation) color space from the RGB color space, and then the histogram equalization process is performed on the V component.
104:将人脸样本数据中的每张所述第一人脸图片输入第一深度神经网络中提取出第一人脸特征,同时将对应的直方图均衡化人脸图片输入第二深度神经网络中提取出第二人脸特征。104: Input each first face image in the face sample data into the first depth neural network to extract the first face feature, and input the corresponding histogram equalized face image into the second depth neural network. The second face feature is extracted.
本较佳实施例中,所述第一深度神经网络与所述第二深度神经网路为具有相同层数(例如,200多层)的神经网络。可选地,所述第一深度神经网络与所述第二深度神经网路采用基于感知机的Inception-Resnet-V2(Inception-V2和Resnet相结合的神经网络)神经网络。所述基于感知机的Inception-Resnet-V2神经网络为现有技术,本申请在此不再赘述。In the preferred embodiment, the first deep neural network and the second deep neural network are neural networks having the same number of layers (for example, 200 layers). Optionally, the first depth neural network and the second deep neural network adopt a perceptron-based Inception-Resnet-V2 (Inception-V2 and Resnet combined neural network) neural network. The perceptron-based Inception-Resnet-V2 neural network is a prior art, and the present application does not repeat here.
优选地,所述步骤104还可以包括:所述第一深度神经网络提取出的所述第一人脸特征的维数为64维,所述第二深度神经网络提取出的所述第二人脸特征的维数为64维。Preferably, the
105:对所述第一人脸特征进行归一化得到第一归一化人脸特征,同时对所述第二人脸特征进行归一化得到第二归一化人脸特征。105: normalize the first facial feature to obtain a first normalized facial feature, and normalize the second facial feature to obtain a second normalized facial feature.
本较佳实施例中,可以采用归一化函数对所述第一人脸特征及所述第二人脸特征进行归一化,所述归一化函数可以是欧式距离函数,也可以是曼哈顿距离函数、最小绝对误差。可选地,采用欧式距离函数对所述第一人脸特征得到第一归一化人脸特征,及对所述第二人脸特征进行归一化得到第二归一化人脸特征。In the preferred embodiment, the first facial feature and the second facial feature may be normalized by a normalization function, and the normalization function may be a Euclidean distance function or a Manhattan Distance function, minimum absolute error. Optionally, the first normalized face feature is obtained by using the Euclidean distance function on the first facial feature, and the second facial feature is normalized to obtain a second normalized facial feature.
归一化能够对光照、阴影和边缘进行压缩,使得第一归一化人脸特征和第二归一化人脸特征对光照、阴影和边缘变化具有鲁棒性,从而进一步提高 双通道神经网络模型的鲁棒性。另外,采用欧式距离函数进行归一化,能避免双通道神经网络模型过拟合,进而提升了双通道神经网络模型的泛化能力,对后续双通道神经网络模型的权重和偏置的优化求解变得更稳定和快速。Normalization can compress light, shadows, and edges so that the first normalized face feature and the second normalized face feature are robust to illumination, shadow, and edge changes, further enhancing the two-channel neural network. Model robustness. In addition, normalization using the Euclidean distance function can avoid over-fitting of the two-channel neural network model, thereby improving the generalization ability of the two-channel neural network model, and optimizing the weight and offset of the subsequent two-channel neural network model. Become more stable and fast.
106:根据预设连接规则连接所述第一归一化人脸特征和所述第二归一化人脸特征得到新的归一化人脸特征,作为最终人脸特征表示。106: Connect the first normalized face feature and the second normalized face feature according to a preset connection rule to obtain a new normalized face feature as a final face feature representation.
本较佳实施例中,所述预设连接规则可以是交叉连接。例如,第一归一化人脸特征为(x 1,x 2,x 3,…,x m),第二归一化人脸特征为(y 1,y 2,y 3,…,y m),则采用交叉连接方式连接得到的最终人脸特征表示为(x 1,y 1,x 2,y 2,x 3,y 3,…,x m,y m)。 In the preferred embodiment, the preset connection rule may be a cross connection. For example, the first normalized face feature is (x 1 , x 2 , x 3 , . . . , x m ), and the second normalized face feature is (y 1 , y 2 , y 3 , . . . , y m ), the final face features obtained by the cross-connection connection are expressed as (x 1 , y 1 , x 2 , y 2 , x 3 , y 3 , . . . , x m , y m ).
可替换地,所述预设连接规则可以是顺序连接。例如,第一归一化人脸特征为(x 1,x 2,x 3,…,x m),第二归一化人脸特征为(y 1,y 2,y 3,…,y m),则采用顺序连接方式连接得到的最终人脸特征表示为(x 1,x 2,x 3,…,x m,y 1,y 2,y 3,…,y m)。 Alternatively, the preset connection rules may be sequential connections. For example, the first normalized face feature is (x 1 , x 2 , x 3 , . . . , x m ), and the second normalized face feature is (y 1 , y 2 , y 3 , . . . , y m ), the final face features obtained by sequential connection are expressed as (x 1 , x 2 , x 3 , . . . , x m , y 1 , y 2 , y 3 , . . . , y m ).
107:将所述最终人脸特征表示进行归一化后输入至预先设置的损失函数中计算损失函数值,当损失函数值小于或等于预先设置的损失函数阈值时,则双通道神经网络模型训练结束并更新所述双通道神经网络模型中的权重和偏置;当损失函数值大于预先设置的损失函数阈值,则获取更多的原始人脸图片,并基于增加后的原始人脸图片重新执行上述步骤101-107。107: normalize the final facial feature representation and input to a preset loss function to calculate a loss function value. When the loss function value is less than or equal to a preset loss function threshold, the dual channel neural network model training Ending and updating the weights and offsets in the two-channel neural network model; when the loss function value is greater than a preset loss function threshold, acquiring more original face images and re-executing based on the increased original face images The above steps 101-107.
本较佳实施例中,所述预先设置的损失函数L为下式(1-3)所示:In the preferred embodiment, the preset loss function L is represented by the following formula (1-3):
其中,m表示mini-batch的大小。L S表示的是交叉熵损失函数(softmax loss)的结果, 代表的是第yi个人的实际输出,L C表示的是中心损失函数(center loss)的结果,cyi表示第yi个人的特征中心,xi表示全连接层之前的特征。预先设置的损失函数L越小,代表模型越好。 Where m is the size of the mini-batch. L S represents the result of the cross-entropy loss function (softmax loss). Represents the actual output of the yi person, L C represents the result of the center loss function, cyi represents the feature center of the yi person, and xi represents the feature before the fully connected layer. The smaller the loss function L set in advance, the better the representative model.
对所述最终人脸特征表示再次进行归一化,是为了使得归一化后的最终人脸特征表示分布在0-1之间,具有统一的批判标准,不仅可以方便后续的数据处理,还能使得损失函数的收敛加快;其次,预先设置的损失函数能够使得不同人脸图片间的距离变大,同一个人的不同人脸图片间的距离减少,从而提高所述双通道神经网络模型的分类能力,进而提高人脸识别的准确度。Normalizing the final facial feature representation again, in order to make the normalized final facial feature representation distributed between 0-1, with a unified critical standard, which not only facilitates subsequent data processing, but also The convergence of the loss function can be accelerated. Secondly, the preset loss function can make the distance between different face images larger, and the distance between different face images of the same person is reduced, thereby improving the classification of the two-channel neural network model. Ability to improve the accuracy of face recognition.
综上所述,本申请所训练出的双通道神经网络模型,能够显著增强对输入的人脸样本数据集中的人脸图片及经过直方图均衡化处理的所述人脸图片的特征提取能力,减轻网络层次加深时而性能降低的风险,提高了双通道神经网络模型的鲁棒性,便于后续提高人脸识别或人脸比对的识别率。In summary, the dual-channel neural network model trained in the present application can significantly enhance the feature extraction capability of the face image in the input face sample data set and the face image processed by the histogram equalization process. It can reduce the risk of performance degradation when the network level is deepened, and improve the robustness of the two-channel neural network model, which is convenient for improving the recognition rate of face recognition or face comparison.
实施例二Embodiment 2
图2是本申请实施例二提供的双通道神经网络模型训练方法的流程图。2 is a flow chart of a two-channel neural network model training method provided by Embodiment 2 of the present application.
201:获取不同人的一张原始人脸图片,每张原始人脸图片的尺寸为182*182。201: Obtain a raw face picture of different people, and the size of each original face picture is 182*182.
本实施例中的所述步骤201同实施例一中的所述步骤101,本申请在此 不再详细赘述。The
202:将每张原始人脸图片进行随机裁剪成预设数量的第二人脸图片,得到人脸样本数据集,其中,所述人脸样本数据集中的每张第二人脸图片的尺寸为171*160。202: Randomly crop each original face image into a preset number of second face images to obtain a face sample data set, where a size of each second face image in the face sample data set is 171*160.
本较佳实施例中,设置随机裁剪后得到的第二人脸图片尺寸为171*160,不仅能最大程度的保持原始人脸图片中的面部细节信息,相较于利用所述原始人脸图片作为人脸样本数据集参与训练时,能缩短双通道神经网络模型的训练时间,提高双通道神经网络模型的训练速度。In the preferred embodiment, the second face image size obtained after random cropping is 171*160, which not only can maintain the facial detail information in the original face image to the greatest extent, but also utilizes the original face image. When the face sample data set is involved in training, the training time of the two-channel neural network model can be shortened, and the training speed of the two-channel neural network model can be improved.
203:利用预先存储的人脸检测算法分别对裁剪出的每张第二人脸图片进行检测,判断裁剪出的所有第二人脸图片中的每张第二人脸图片中是否包含人脸区域。203: Detect each second face image that is cropped by using a pre-stored face detection algorithm, and determine whether each second face image in the cropped second face image includes a face region. .
所述预先存储的人脸检测算法可以为以下算法中的一种:基于模板的人脸检测方法、基于人工神经网络的人脸检测方法、基于模型的人脸检测方法、基于肤色的人脸检测方法或者基于特征子脸的人脸检测方法等。The pre-stored face detection algorithm may be one of the following algorithms: a template-based face detection method, an artificial neural network-based face detection method, a model-based face detection method, and a skin color-based face detection method. The method or the face detection method based on the feature sub-face, and the like.
本实施例中,所述预先存储的人脸检测算法不限于上述列举的,任何能够检测裁剪出的人脸图片中是否包含人脸区域的算法均可引用与此。另外,本实施例中预先存储的人脸检测算法为现有技术,本文在此不再详细介绍。In this embodiment, the pre-stored face detection algorithm is not limited to the above enumeration, and any algorithm capable of detecting whether the face image is included in the cropped face image may be referred to herein. In addition, the face detection algorithm pre-stored in this embodiment is a prior art, and will not be described in detail herein.
当确定裁剪出的所有第二人脸图片中的每张第二人脸图片中包含人脸区域时,则执行下述步骤204;当确定裁剪出的所有第二人脸图片中的每张或部分人脸图片中不包含人脸区域时,则执行下述步骤205。When it is determined that each of the second face images in the cropped second face image includes a face region, then performing the following
所述裁剪出的第二人脸图片中不包含人脸区域包括以下两种情形:完全不存在人脸区域;仅存在部分人脸区域。The face region that is not included in the cropped second face image includes the following two situations: there is no face region at all; only a part of the face region exists.
204:分别对所述人脸样本数据集中的每张第二人脸图片进行直方图均衡化处理,得到直方图均衡化人脸图片。204: Perform a histogram equalization process on each second face image in the face sample data set to obtain a histogram equalization face image.
本实施例中的所述步骤204同实施例一中的所述步骤103,本申请在此不再详细赘述。The
205:继续对不包含人脸区域的第二人脸图片所对应的原始人脸图片进行随机裁剪,直到得到所述预设数量的第二人脸图片。205: Continue to perform random cropping on the original face image corresponding to the second face image that does not include the face region, until the preset number of second face images are obtained.
206:将人脸样本数据中的每张所述第二人脸图片输入第一深度神经网络中提取出第一人脸特征,同时将对应的直方图均衡化人脸图片输入第二深度神经网络中提取出第二人脸特征。206: Input each second face image in the face sample data into the first depth neural network to extract the first face feature, and input the corresponding histogram equalized face image into the second depth neural network. The second face feature is extracted.
本实施例中的所述步骤206同实施例一中的所述步骤104,本申请在此不再详细赘述。The
207:对所述第一人脸特征进行归一化得到第一归一化人脸特征,同时对所述第二人脸特征进行归一化得到第二归一化人脸特征。207: normalize the first facial feature to obtain a first normalized facial feature, and normalize the second facial feature to obtain a second normalized facial feature.
本实施例中的所述步骤207同实施例一中的所述步骤105,本申请在此不再详细赘述。The
208:根据预设连接规则连接所述第一归一化人脸特征和所述第二归一化人脸特征得到新的归一化人脸特征,作为最终人脸特征表示。208: Connect the first normalized face feature and the second normalized face feature according to a preset connection rule to obtain a new normalized face feature as a final face feature representation.
本实施例中的所述步骤208同实施例一中的所述步骤106,本申请在此 不再详细赘述。The
209:将所述最终人脸特征表示进行归一化后输入至预先设置的损失函数中计算损失函数值,当损失函数值小于或等于预先设置的损失函数阈值时,则双通道神经网络模型训练结束并更新所述双通道神经网络模型中的权重和偏置;当损失函数值大于预先设置的损失函数阈值,则获取更多的原始人脸图片,并基于增加后的原始人脸图片重新执行上述步骤201-209。209: normalize the final facial feature representation and input to a preset loss function to calculate a loss function value. When the loss function value is less than or equal to a preset loss function threshold, the dual channel neural network model training Ending and updating the weights and offsets in the two-channel neural network model; when the loss function value is greater than a preset loss function threshold, acquiring more original face images and re-executing based on the increased original face images The above steps 201-209.
本实施例中的所述步骤209同实施例一中的所述步骤107,本申请在此不再详细赘述。The
综上所述,本申请所训练出的双通道神经网络模型,通过预先存储的人脸检测算法对裁剪出的每张第二人脸图片进行检测,以确保每张第二人脸图片中包含了全部的人脸区域,如此可以保障提取的人脸特征的有效性,能够进一步提高双通道神经网络模型的鲁棒性,便于后续提高人脸识别或人脸比对的识别率。In summary, the two-channel neural network model trained in the present application detects each of the cropped second face images by a pre-stored face detection algorithm to ensure that each second face image is included. All the face regions can ensure the validity of the extracted face features, which can further improve the robustness of the two-channel neural network model, and facilitate the subsequent recognition of face recognition or face comparison.
实施例三Embodiment 3
图3是本申请实施例三提供的人脸比对方法的流程图。FIG. 3 is a flowchart of a face matching method according to Embodiment 3 of the present application.
301:在收到待进行人脸比对的目标用户的人脸图片后,对目标用户的人脸图片进行直方图均衡化处理,得到目标用户的直方图均衡化人脸图片。301: After receiving the face image of the target user to be compared, perform a histogram equalization process on the face image of the target user to obtain a histogram equalized face image of the target user.
302:将所述目标用户的人脸图片输入训练好的双通道神经网络模型中的第一深度神经网络中提取出第三人脸特征,同时将所述目标用户的直方图均衡化人脸图片输入训练好的双通道神经网络模型中的第二深度神经网络中提取出第四人脸特征。302: Enter a face image of the target user into a first deep neural network in the trained two-channel neural network model to extract a third facial feature, and simultaneously equalize the face image of the target user The fourth facial feature is extracted from the second deep neural network in the trained two-channel neural network model.
303:对所述第三人脸特征进行归一化得到第三归一化人脸特征,同时对所述第四人脸特征进行归一化得到第四归一化人脸特征。303: normalize the third facial feature to obtain a third normalized facial feature, and normalize the fourth facial feature to obtain a fourth normalized facial feature.
304:根据所述预设连接规则连接所述第三归一化人脸特征和所述第四归一化人脸特征得到新的归一化人脸特征,作为所述目标用户的最终特征表示。304: Connect the third normalized face feature and the fourth normalized face feature according to the preset connection rule to obtain a new normalized face feature, as a final feature representation of the target user. .
305:计算所述目标用户的最终特征表示与数据库中注册用户的最终特征表示之间的相似度。305: Calculate a similarity between the final feature representation of the target user and the final feature representation of the registered user in the database.
根据式(2-1)和(2-2)计算所述目标用户的最终特征表示与数据库中注册用户的最终特征表示之间的相似度。The similarity between the final feature representation of the target user and the final feature representation of the registered user in the database is calculated according to equations (2-1) and (2-2).
SIM=0.5+0.5*COSθ (2-2)SIM=0.5+0.5*COSθ (2-2)
其中,特征向量A代表目标用户的最终特征表示,特征向量B代表注册用户的最终特征表示。根据式(2-1)计算特征向量A和数据库中的每一个特征向量B之间的余弦值,计算出的余弦值的范围为[-1,1]。根据式(2-2)计算出的SIM即为相似度,此时相似度的值归一化到[0,1]。Wherein, the feature vector A represents the final feature representation of the target user, and the feature vector B represents the final feature representation of the registered user. The cosine value between the feature vector A and each of the feature vectors B in the database is calculated according to the equation (2-1), and the range of the calculated cosine value is [-1, 1]. The SIM calculated according to equation (2-2) is the similarity, and the value of the similarity is normalized to [0, 1].
306:根据计算得到的相似度与预先设置的置信度阈值之间的关系,确定 人脸比对结果。306: Determine a face comparison result according to a relationship between the calculated similarity and a preset confidence threshold.
所述人脸比对的结果包括比对通过结果和比对失败结果。当计算得到的相似度大于或等于预先设置的置信度阈值时,则确定用户比对通过,即在数据库中匹配出了与目标用户为同一个人的已注册用户。当计算得到的相似度小于预先设置的置信度阈值时,则确定用户比对失败,即在数据库中没有匹配出与目标用户为同一个人的已注册用户。The results of the face alignment include alignment results and comparison failure results. When the calculated similarity is greater than or equal to the preset confidence threshold, it is determined that the user matches, that is, the registered user who is the same person as the target user is matched in the database. When the calculated similarity is less than the preset confidence threshold, it is determined that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
优选地,所述预先设置的置信度阈值为0.5。Preferably, the preset confidence threshold is 0.5.
综上所述,将所述双通道神经网络模型用于人脸比对,能够提高人脸比对的准确度,缩短人脸比对的时间,提高人脸比对的效率。In summary, the dual-channel neural network model is used for face comparison, which can improve the accuracy of face comparison, shorten the time of face comparison, and improve the efficiency of face comparison.
实施例四Embodiment 4
图4是本申请实施例四提供的人脸比对方法的流程图。4 is a flowchart of a face matching method provided in Embodiment 4 of the present application.
401:在收到待进行人脸比对的目标用户的人脸图片后,对所述目标用户的人脸图片进行预处理。401: After receiving the face image of the target user to be compared with the face, pre-processing the face image of the target user.
本较佳实施例中,所述预处理可以包括,但不限于:图像去噪,光照归一化,姿态校准,灰度归一化等。In the preferred embodiment, the pre-processing may include, but is not limited to, image denoising, illumination normalization, pose calibration, grayscale normalization, and the like.
在收到待进行人脸比对的目标用户的人脸图片后,对所述目标用户的人脸图片进行预处理,是为了进一步提高目标用户的人脸图片,便于提取出更具辨别力的人脸特征。After receiving the face image of the target user to be compared, the face image of the target user is pre-processed to further improve the face image of the target user, so as to extract more discriminative power. Face features.
402:对经过预处理后目标用户的人脸图片进行直方图均衡化处理,得到目标用户的直方图均衡化人脸图片。402: Perform a histogram equalization process on the face image of the target user after the pre-processing, and obtain a histogram equalization face image of the target user.
403:将经过预处理后的目标用户的人脸图片输入训练好的双通道神经网络模型中的第一深度神经网络中提取出第三人脸特征,同时将所述目标用户的直方图均衡化人脸图片输入训练好的双通道神经网络模型中的第二深度神经网络中提取出第四人脸特征。403: Input the pre-processed target user's face image into the first deep neural network in the trained two-channel neural network model to extract the third facial feature, and simultaneously equalize the target user's histogram. The fourth face feature is extracted from the second deep neural network in the trained two-channel neural network model.
404:对所述第三人脸特征进行归一化得到第三归一化人脸特征,同时对所述第四人脸特征进行归一化得到第四归一化人脸特征。404: normalize the third facial feature to obtain a third normalized facial feature, and normalize the fourth facial feature to obtain a fourth normalized facial feature.
405:根据所述预设连接规则连接所述第三归一化人脸特征和所述第四归一化人脸特征得到新的归一化人脸特征,作为所述目标用户的最终特征表示。405: Connect the third normalized face feature and the fourth normalized face feature according to the preset connection rule to obtain a new normalized face feature, as a final feature representation of the target user. .
406:计算所述目标用户的最终特征表示与数据库中注册用户的最终特征表示之间的相似度。406: Calculate a similarity between the final feature representation of the target user and the final feature representation of the registered user in the database.
407:根据预先设置的排序方法对计算得到的相似度进行排序。407: Sort the calculated similarities according to a preset sorting method.
本较佳实施例中,所述预先设置的排序方法包括,但不限于:由大到小排序,由小到大排序。In the preferred embodiment, the pre-set sorting method includes, but is not limited to, sorting from large to small, and sorting from small to large.
408:比较最大的相似度与预先设置的置信度阈值之间的关系,确定人脸比对结果。408: Compare the relationship between the maximum similarity and the preset confidence threshold to determine the face comparison result.
所述人脸比对的结果包括比对通过结果和比对失败结果。当最大的相似度大于或等于预先设置的置信度阈值时,则确定用户比对通过,即在数据库中匹配出了与目标用户为同一个人的已注册用户。当最大的相似度小于预先设置的置信度阈值时,则确定用户比对失败,即在数据库中没有匹配出与目标用户为同一个人的已注册用户。The results of the face alignment include alignment results and comparison failure results. When the maximum similarity is greater than or equal to the preset confidence threshold, it is determined that the user matches, that is, the registered user who is the same person as the target user is matched in the database. When the maximum similarity is less than the preset confidence threshold, it is determined that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
优选地,所述预先设置的置信度阈值为0.5。Preferably, the preset confidence threshold is 0.5.
通过对计算得到的相似度进行排序,并直接比较最大的相似度与预先设置的置信度阈值之间的关系,当最大的相似度大于或等于预先设置的置信度阈值时,则确定用户比对通过。当最大的相似度小于预先设置的置信度阈值时,则确定用户比对失败。如此可以缩短比对时间,仅需比对一次即可确定目标用户是否为已注册用户。By sorting the calculated similarities and directly comparing the relationship between the maximum similarity and the preset confidence threshold, when the maximum similarity is greater than or equal to the preset confidence threshold, the user comparison is determined. by. When the maximum similarity is less than the preset confidence threshold, it is determined that the user comparison fails. This can shorten the comparison time, and only need to compare once to determine whether the target user is a registered user.
综上所述,本申请所述的人脸比对方法,使用双通道神经网络模型能够提取出较佳的人脸特征表示,花费较短的人脸比对时间即可获得较佳的人脸比对效果。In summary, the face matching method described in the present application can extract a better facial feature representation by using a two-channel neural network model, and a better face can be obtained by taking a shorter face comparison time. Compare effects.
上述对图1、图2、图3和图4的流程图中各步骤的描述,根据不同的需求,流程图中的执行顺序可以改变,某些步骤可以省略。The above description of the steps in the flowcharts of FIG. 1, FIG. 2, FIG. 3 and FIG. 4 may be performed according to different requirements, and the order of execution in the flowchart may be changed, and some steps may be omitted.
以上所述,仅是本申请的具体实施方式,但本申请的保护范围并不局限于此,对于本领域的普通技术人员来说,在不脱离本申请创造构思的前提下,还可以做出改进,但这些均属于本申请的保护范围。The above description is only a specific embodiment of the present application, but the scope of protection of the present application is not limited thereto, and those skilled in the art can also make without departing from the concept of the present application. Improvements, but these are all within the scope of this application.
下面结合第5至7图,分别对实现上述双通道神经网络模型训练方法及人脸比对方法的终端的功能模块及硬件结构进行介绍。The function modules and hardware structures of the terminal for realizing the above-described two-channel neural network model training method and face comparison method are respectively described below with reference to the fifth to seventh figures.
实施例五Embodiment 5
图5为本申请双通道神经网络模型训练装置较佳实施例中功能模块图。FIG. 5 is a functional block diagram of a preferred embodiment of the dual channel neural network model training device of the present application.
在一些实施例中,所述双通道神经网络模型训练装置50运行于所述终端7中。所述双通道神经网络模型训练装置50可以包括多个由计算机可读指令代码段所组成的功能模块。所述双通道神经网络模型训练装置50中的各个计算机可读指令代码段可以存储于所述存储器71中,并由所述至少一个处理器72所执行,以执行(详见图1和图2描述)对双通道神经网络模型的训练。In some embodiments, the two-channel neural network model training device 50 operates in the
本实施例中,所述双通道神经网络模型训练装置50根据其所执行的功能,可以被划分为多个功能模块。所述功能模块可以包括:获取模块501、裁剪模块502、第一直方图处理模块503、第一特征提取模块504、第一归一化模块505、第一特征连接模块506、模型更新模块507及人脸检测模块508。In this embodiment, the dual channel neural network model training device 50 can be divided into multiple functional modules according to the functions performed by the dual channel neural network model. The function module may include: an obtaining
获取模块501,用于获取不同人的一张原始人脸图片,每张原始人脸图片的尺寸为182*182。The obtaining
裁剪模块502,用于将每张原始人脸图片进行随机裁剪成预设数量的第一人脸图片,得到人脸样本数据集。The
所述预设数量可以根据实际需要自行调整或修改,本较佳实施例中,所述裁剪模块502将每张原始人脸图片进行随机裁剪成大小相同的5张第一人脸图片,每张第一人脸图片均包含人脸区域。The preset number can be adjusted or modified according to actual needs. In the preferred embodiment, the
将每张原始人脸图片进行随机裁剪,可得到任意角度、任意光照等条件下的第一人脸图片,随机的过程即是模拟动态环境下采集原始人脸图片的过程;其次,通过裁剪的方式可得到多张第一人脸图片,从而增加了人脸样本数据集的数量,较多的人脸样本数据集能够提高双通道神经网络模型的鲁棒性。Randomly crop each original face image to obtain the first face image under arbitrary angle and arbitrary illumination conditions. The random process is the process of collecting the original face image in the simulated dynamic environment; secondly, by cutting In this way, multiple first face images can be obtained, thereby increasing the number of face sample data sets, and more face sample data sets can improve the robustness of the two-channel neural network model.
第一直方图处理模块503,用于分别对所述人脸样本数据集中的每张第 一人脸图片进行直方图均衡化处理,得到直方图均衡化人脸图片。The
第一特征提取模块504,用于将人脸样本数据中的每张所述第一人脸图片输入第一深度神经网络中提取出第一人脸特征,同时将对应的直方图均衡化人脸图片输入第二深度神经网络中提取出第二人脸特征。The first feature extraction module 504 is configured to input each first face image in the face sample data into the first depth neural network to extract the first facial feature, and simultaneously equalize the corresponding histogram. The picture input second depth neural network extracts the second face feature.
本较佳实施例中,所述第一深度神经网络与所述第二深度神经网路为具有相同层数(例如,200多层)的神经网络。可选地,所述第一深度神经网络与所述第二深度神经网路采用基于感知机的Inception-Resnet-V2(Inception-V2和Resnet相结合的神经网络)神经网络。所述基于感知机的Inception-Resnet-V2神经网络为现有技术,本申请在此不再赘述。In the preferred embodiment, the first deep neural network and the second deep neural network are neural networks having the same number of layers (for example, 200 layers). Optionally, the first depth neural network and the second deep neural network adopt a perceptron-based Inception-Resnet-V2 (Inception-V2 and Resnet combined neural network) neural network. The perceptron-based Inception-Resnet-V2 neural network is a prior art, and the present application does not repeat here.
优选地,所述第一特征提取模块504,还用于所述第一深度神经网络提取出的所述第一人脸特征的维数为64维,所述第二深度神经网络提取出的所述第二人脸特征的维数为64维。Preferably, the first feature extraction module 504 is further configured to: the dimension of the first facial feature extracted by the first depth neural network is 64 dimensions, and the second depth neural network extracts The dimension of the second face feature is 64 dimensions.
第一归一化模块505,用于对所述第一人脸特征进行归一化得到第一归一化人脸特征,同时对所述第二人脸特征进行归一化得到第二归一化人脸特征。a first normalization module 505, configured to normalize the first facial feature to obtain a first normalized facial feature, and normalize the second facial feature to obtain a second normalized Face features.
第一特征连接模块506,用于根据预设连接规则连接所述第一归一化人脸特征和所述第二归一化人脸特征得到新的归一化人脸特征,作为最终人脸特征表示。a first feature connection module 506, configured to connect the first normalized face feature and the second normalized face feature according to a preset connection rule to obtain a new normalized face feature, as a final face Feature representation.
模型更新模块507,用于将所述最终人脸特征表示进行归一化后输入至预先设置的损失函数中计算损失函数值,当损失函数值小于或等于预先设置的损失函数阈值时,则双通道神经网络模型训练结束并更新所述双通道神经网络模型中的权重和偏置;当损失函数值大于预先设置的损失函数阈值,则获取更多的原始人脸图片,并基于增加后的原始人脸图片重新执行上述模块501-507。The model update module 507 is configured to normalize the final facial feature representation and input the value to the preset loss function to calculate a loss function value. When the loss function value is less than or equal to a preset loss function threshold, the double The channel neural network model training ends and updates the weights and offsets in the two-channel neural network model; when the loss function value is greater than the preset loss function threshold, more original face images are acquired, and based on the increased original The face picture re-executes the above modules 501-507.
进一步地,所述裁剪模块502还用于将每张原始人脸图片进行随机裁剪成预设数量的第二人脸图片,得到人脸样本数据集,其中,所述人脸样本数据集中的每张第二人脸图片的尺寸为171*160。Further, the
本较佳实施例中,设置随机裁剪后得到的第二人脸图片尺寸为171*160,不仅能最大程度的保持原始人脸图片中的面部细节信息,相较于利用所述原始人脸图片作为人脸样本数据集参与训练时,能缩短双通道神经网络模型的训练时间,提高双通道神经网络模型的训练速度。In the preferred embodiment, the second face image size obtained after random cropping is 171*160, which not only can maintain the facial detail information in the original face image to the greatest extent, but also utilizes the original face image. When the face sample data set is involved in training, the training time of the two-channel neural network model can be shortened, and the training speed of the two-channel neural network model can be improved.
进一步地,所述双通道神经网路模型训练装置50还包括:人脸检测模块508,用于利用预先存储的人脸检测算法分别对裁剪出的每张第二人脸图片进行检测,判断裁剪出的所有第二人脸图片中的每张第二人脸图片中是否包含人脸区域。Further, the dual-channel neural network model training device 50 further includes: a face detection module 508, configured to detect each of the cropped second face images by using a pre-stored face detection algorithm, and determine the cropping Whether or not the face area is included in each second face picture in all the second face pictures.
所述预先存储的人脸检测算法可以为以下算法中的一种:基于模板的人脸检测方法、基于人工神经网络的人脸检测方法、基于模型的人脸检测方法、基于肤色的人脸检测方法或者基于特征子脸的人脸检测方法等。The pre-stored face detection algorithm may be one of the following algorithms: a template-based face detection method, an artificial neural network-based face detection method, a model-based face detection method, and a skin color-based face detection method. The method or the face detection method based on the feature sub-face, and the like.
本实施例中,所述预先存储的人脸检测算法不限于上述列举的,任何能够检测裁剪出的人脸图片中是否包含人脸区域的算法均可引用与此。另外, 本实施例中预先存储的人脸检测算法为现有技术,本文在此不再详细介绍。In this embodiment, the pre-stored face detection algorithm is not limited to the above enumeration, and any algorithm capable of detecting whether the face image is included in the cropped face image may be referred to herein. In addition, the face detection algorithm pre-stored in this embodiment is a prior art, and will not be described in detail herein.
进一步地,所述第一直方图处理模块503还用于当所述人脸检测模块508确定裁剪出的所有第二人脸图片中的每张第二人脸图片中包含人脸区域时,分别对所述人脸样本数据集中的每张第二人脸图片进行直方图均衡化处理,得到直方图均衡化人脸图片。Further, the first
进一步地,所述裁剪模块502还用于当所述人脸检测模块508确定裁剪出的所有第二人脸图片中的每张或部分人脸图片中不包含人脸区域时,继续对不包含人脸区域的第二人脸图片所对应的原始人脸图片进行随机裁剪,直到得到所述预设数量的第二人脸图片。Further, the
所述裁剪出的第二人脸图片中不包含人脸区域包括以下两种情形:完全不存在人脸区域;仅存在部分人脸区域。The face region that is not included in the cropped second face image includes the following two situations: there is no face region at all; only a part of the face region exists.
进一步地,所述第一归一化模块505还用于将人脸样本数据中的每张所述第二人脸图片输入第一深度神经网络中提取出第一人脸特征,同时将对应的直方图均衡化人脸图片输入第二深度神经网络中提取出第二人脸特征。Further, the first normalization module 505 is further configured to input each second face image in the face sample data into the first depth neural network to extract the first facial feature, and correspondingly The histogram equalization face image input second depth neural network extracts the second face feature.
综上所述,本申请所训练出的双通道神经网络模型,能够显著增强对输入的人脸样本数据集中的人脸图片及经过直方图均衡化处理的所述人脸图片的特征提取能力,减轻网络层次加深时而性能降低的风险,提高了双通道神经网络模型的鲁棒性;其次,通过预先存储的人脸检测算法对裁剪出的每张第二人脸图片进行检测,以确保每张第二人脸图片中包含了全部的人脸区域,如此可以保障提取的人脸特征的有效性,能够进一步提高双通道神经网络模型的鲁棒性,便于后续提高人脸识别或人脸比对的识别率。In summary, the dual-channel neural network model trained in the present application can significantly enhance the feature extraction capability of the face image in the input face sample data set and the face image processed by the histogram equalization process. It mitigates the risk of performance degradation when the network level is deepened, and improves the robustness of the two-channel neural network model. Secondly, each second face image that is cropped is detected by a pre-stored face detection algorithm to ensure each piece. The second face image contains all the face regions, which can ensure the validity of the extracted face features, and can further improve the robustness of the two-channel neural network model, and facilitate the subsequent improvement of face recognition or face comparison. Recognition rate.
实施例六Embodiment 6
图6为本申请人脸比对装置的较佳实施例中的功能模块图。FIG. 6 is a functional block diagram of a preferred embodiment of the applicant's face matching device.
在一些实施例中,所述人脸比对装置60运行于所述终端7中。所述人脸比对装置60可以包括多个由计算机可读指令代码段所组成的功能模块。所述人脸比对装置60中的各个计算机可读指令段的指令代码可以存储于所述存储器71中,并由所述至少一个处理器72所执行,以执行(详见图3和图4描述)对人脸的比对。In some embodiments, the face matching device 60 operates in the
本实施例中,所述终端7的人脸比对装置60根据其所执行的功能,可以被划分为多个功能模块。所述功能模块可以包括:第二直方图处理模块601、第二特征提取模块602、第二归一化模块603、第二特征连接模块604、计算模块605、结果确定模块606、预处理模块607及排序模块608。In this embodiment, the face matching device 60 of the
第二直方图处理模块601,用于在收到待进行人脸比对的目标用户的人脸图片后,对目标用户的人脸图片进行直方图均衡化处理,得到目标用户的直方图均衡化人脸图片。The second
第二特征提取模块602,用于将所述目标用户的人脸图片输入训练好的双通道神经网络模型中的第一深度神经网络中提取出第三人脸特征,同时将所述目标用户的直方图均衡化人脸图片输入训练好的双通道神经网络模型中的第二深度神经网络中提取出第四人脸特征;a second
第二归一化模块603,用于对第三人脸特征进行归一化得到第三归一化 人脸特征,同时对第四人脸特征进行归一化得到第四归一化人脸特征;The second normalization module 603 is configured to normalize the third facial feature to obtain a third normalized facial feature, and normalize the fourth facial feature to obtain a fourth normalized facial feature. ;
第二特征连接模块604,用于根据所述预设连接规则连接所述第三归一化人脸特征和所述第四归一化人脸特征得到新的归一化人脸特征,作为所述目标用户的最终特征表示。The second feature connection module 604 is configured to connect the third normalized face feature and the fourth normalized face feature according to the preset connection rule to obtain a new normalized face feature. The final feature representation of the target user.
计算模块605,用于计算所述目标用户的最终特征表示与数据库中注册用户的最终特征表示之间的相似度。The calculating
结果确定模块606,用于根据计算得到的相似度与预先设置的置信度阈值之间的关系,确定人脸比对结果。The result determining module 606 is configured to determine a face comparison result according to a relationship between the calculated similarity and a preset confidence threshold.
所述人脸比对的结果包括比对通过结果和比对失败结果。当计算得到的相似度大于或等于预先设置的置信度阈值时,则所述结果确定模块606确定用户比对通过,即在数据库中匹配出了与目标用户为同一个人的已注册用户。当计算得到的相似度小于预先设置的置信度阈值时,则所述结果确定模块606确定用户比对失败,即在数据库中没有匹配出与目标用户为同一个人的已注册用户。The results of the face alignment include alignment results and comparison failure results. When the calculated similarity is greater than or equal to the preset confidence threshold, the result determination module 606 determines that the user matches, that is, the registered user who is the same person as the target user is matched in the database. When the calculated similarity is less than the preset confidence threshold, the result determination module 606 determines that the user comparison fails, that is, the registered user who is the same person as the target user is not matched in the database.
优选地,所述预先设置的置信度阈值为0.5。Preferably, the preset confidence threshold is 0.5.
优选地,所述人脸比对装置60还包括:预处理模块607,用于在收到待进行人脸比对的目标用户的人脸图片后,对所述目标用户的人脸图片进行预处理。本较佳实施例中,所述预处理可以包括,但不限于:图像去噪,光照归一化,姿态校准,灰度归一化等。Preferably, the face matching device 60 further includes: a pre-processing module 607, configured to pre-process the face image of the target user after receiving the face image of the target user to be compared by the face deal with. In the preferred embodiment, the pre-processing may include, but is not limited to, image denoising, illumination normalization, pose calibration, grayscale normalization, and the like.
在收到待进行人脸比对的目标用户的人脸图片后,对所述目标用户的人脸图片进行预处理,是为了进一步提高目标用户的人脸图片,便于提取出更具辨别力的人脸特征。After receiving the face image of the target user to be compared, the face image of the target user is pre-processed to further improve the face image of the target user, so as to extract more discriminative power. Face features.
进一步的,所述第二直方图处理模块601还对经过预处理后目标用户的人脸图片进行直方图均衡化处理,得到目标用户的直方图均衡化人脸图片。Further, the second
进一步的,所述第二特征提取模块602还用于将经过预处理后的目标用户的人脸图片输入训练好的双通道神经网络模型中的第一深度神经网络中提取出第三人脸特征,同时将所述目标用户的直方图均衡化人脸图片输入训练好的双通道神经网络模型中的第二深度神经网络中提取出第四人脸特征;Further, the second
优选地,所述人脸比对装置60还包括:排序模块608,用于根据预先设置的排序方法对计算得到的相似度进行排序。本较佳实施例中,所述预先设置的排序方法包括,但不限于:由大到小排序,由小到大排序。Preferably, the face matching device 60 further includes: a sorting
进一步的,所述结果确定模块606还用于比较最大的相似度与预先设置的置信度阈值之间的关系,确定人脸比对结果。Further, the result determining module 606 is further configured to compare a relationship between the maximum similarity and a preset confidence threshold to determine a face comparison result.
所述人脸比对的结果包括比对通过结果和比对失败结果。当最大的相似度大于或等于预先设置的置信度阈值时,则结果确定模块606确定用户比对通过,即在数据库中匹配出了与目标用户为同一个人的已注册用户。当最大的相似度小于预先设置的置信度阈值时,则结果确定模块606确定用户比对失败,即在数据库中没有匹配出与目标用户为同一个人的已注册用户。The results of the face alignment include alignment results and comparison failure results. When the maximum similarity is greater than or equal to the preset confidence threshold, the result determination module 606 determines that the user has passed the comparison, ie, the registered user who is the same person as the target user is matched in the database. When the maximum similarity is less than the preset confidence threshold, the result determination module 606 determines that the user comparison failed, ie, the registered user who is the same person as the target user is not matched in the database.
综上所述,本申请所述的人脸比对方法,使用双通道神经网络模型能够提取出较佳的人脸特征表示,花费较短的人脸比对时间即可获得较佳的人脸 比对效果。In summary, the face matching method described in the present application can extract a better facial feature representation by using a two-channel neural network model, and a better face can be obtained by taking a shorter face comparison time. Compare effects.
实施例七Example 7
如图7所示,是实现本申请所述双通道神经网络模型训练方法及/或所述人脸比对方法的终端的硬件结构示意图。As shown in FIG. 7 , it is a hardware structure diagram of a terminal for implementing the dual channel neural network model training method and/or the face matching method described in the present application.
在本申请较佳实施例中,所述终端7包括存储器71、至少一个处理器72、至少一条通信总线73。In the preferred embodiment of the present application, the
本领域技术人员应该了解,图7示出的终端7的结构并不构成本申请实施例的限定,既可以是总线型结构,也可以是星形结构,所述终端7还可以包括比图示更多或更少的其他硬件或者软件,或者不同的部件布置。It should be understood by those skilled in the art that the structure of the
在一些实施例中,所述存储器71用于存储计算机可读指令代码和各种数据,例如安装在所述终端7中的双通道神经网络模型训练装置50及人脸比对装置60,并在终端7的运行过程中实现高速、自动地完成计算机可读指令或数据的存取。In some embodiments, the
在一些实施例中,所述至少一个处理器72可以由集成电路组成,例如可以由单个封装的集成电路所组成,也可以是由多个相同功能或不同功能封装的集成电路所组成,包括一个或者多个中央处理器、微处理器、数字处理芯片、图形处理器及各种控制芯片的组合等。所述至少一个处理器72是所述终端7的控制核心,利用各种接口和线路连接整个终端7的各个部件,通过运行或执行存储在所述存储器71内的计算机可读指令或者模块,以及调用存储在所述存储器71内的数据,以执行终端7的各种功能和处理数据,例如执行双通道神经网络模型训练装置50及/或人脸比对装置60。In some embodiments, the at least one processor 72 may be composed of an integrated circuit, for example, may be composed of a single packaged integrated circuit, or may be composed of multiple integrated circuits of the same function or different functions, including one. Or a combination of a plurality of central processing units, microprocessors, digital processing chips, graphics processors, and various control chips. The at least one processor 72 is a control core of the
在一些实施例中,所述至少一条通信总线73被设置为实现所述存储器71、所述至少一个处理器72等之间的连接通信。In some embodiments, the at least one
具体地,所述至少一个处理器72对上述指令的具体实现方法可参考图1至图4对应实施例中相关步骤的描述,在此不赘述。For a specific implementation of the above-mentioned instructions by the at least one processor 72, reference may be made to the description of related steps in the corresponding embodiments in FIG. 1 to FIG. 4, and details are not described herein.
在本申请各个实施例中的各功能模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能模块的形式实现。The functional modules in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of hardware plus software function modules.
对于本领域技术人员而言,显然本申请不限于上述示范性实施例的细节,而且在不背离本申请的精神或基本特征的情况下,能够以其他的具体形式实现本申请。此外,显然“包括”一词不排除其他单元或,单数不排除复数。系统权利要求中陈述的多个单元或装置也可以由一个单元或装置通过软件或硬件来实现。第一,第二等词语用来表示名称,而并不表示任何特定的顺序。It is obvious to those skilled in the art that the present application is not limited to the details of the above-described exemplary embodiments, and the present invention can be implemented in other specific forms without departing from the spirit or essential characteristics of the present application. In addition, it is to be understood that the term "comprising" does not exclude other elements or the singular does not exclude the plural. The plurality of units or devices recited in the system claims can also be implemented by a unit or device by software or hardware. The first, second, etc. words are used to denote names and do not denote any particular order.
最后应说明的是,以上实施例仅用以说明本申请的技术方案而非限制,尽管参照较佳实施例对本申请进行了详细说明,本领域的普通技术人员应当理解,可以对本申请的技术方案进行修改或等同替换,而不脱离本申请技术方案的精神和范围。It should be noted that the above embodiments are only used to explain the technical solutions of the present application, and are not limited thereto. Although the present application is described in detail with reference to the preferred embodiments, those skilled in the art should understand that the technical solutions of the present application can be applied. Modifications or equivalents are made without departing from the spirit and scope of the technical solutions of the present application.
Claims (20)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810299180.6A CN108446666A (en) | 2018-04-04 | 2018-04-04 | The training of binary channels neural network model and face comparison method, terminal and medium |
| CN201810299180.6 | 2018-04-04 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2019192121A1 true WO2019192121A1 (en) | 2019-10-10 |
Family
ID=63199349
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2018/100152 Ceased WO2019192121A1 (en) | 2018-04-04 | 2018-08-13 | Dual-channel neural network model training and human face comparison method, and terminal and medium |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN108446666A (en) |
| WO (1) | WO2019192121A1 (en) |
Cited By (22)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110992329A (en) * | 2019-11-28 | 2020-04-10 | 上海微创医疗器械(集团)有限公司 | Product surface defect detection method, electronic device and readable storage medium |
| CN111027421A (en) * | 2019-11-26 | 2020-04-17 | 西安宏规电子科技有限公司 | Graph-based direct-push type semi-supervised pedestrian re-identification method |
| CN111144332A (en) * | 2019-12-30 | 2020-05-12 | 深圳云天励飞技术有限公司 | A kind of picture gathering method, device and electronic device |
| CN111144566A (en) * | 2019-12-30 | 2020-05-12 | 深圳云天励飞技术有限公司 | Neural network weight parameter training method, feature classification method and corresponding device |
| CN111198963A (en) * | 2019-12-11 | 2020-05-26 | 智慧眼科技股份有限公司 | Target retrieval method and device based on average characteristics and related equipment thereof |
| CN111209839A (en) * | 2019-12-31 | 2020-05-29 | 上海涛润医疗科技有限公司 | Face recognition method |
| CN111324589A (en) * | 2020-02-14 | 2020-06-23 | 中国平安人寿保险股份有限公司 | Log sending method and device based on similarity degree value and computer equipment |
| CN111582358A (en) * | 2020-04-30 | 2020-08-25 | 上海添玑网络服务有限公司 | Training method and device for house type recognition model and house type weight judging method and device |
| CN111612785A (en) * | 2020-06-03 | 2020-09-01 | 浙江大华技术股份有限公司 | Method and device for evaluating quality of face picture and storage medium |
| CN111709890A (en) * | 2020-06-12 | 2020-09-25 | 北京小米松果电子有限公司 | An image enhancement model training method, device and storage medium |
| CN112883213A (en) * | 2019-11-29 | 2021-06-01 | 深圳云天励飞技术有限公司 | Picture archiving method and device and electronic equipment |
| CN113657289A (en) * | 2021-08-19 | 2021-11-16 | 北京百度网讯科技有限公司 | Training method, device and electronic device for threshold estimation model |
| CN114004779A (en) * | 2020-07-27 | 2022-02-01 | 中移物联网有限公司 | Face quality evaluation method and device based on deep learning |
| CN114187309A (en) * | 2022-01-11 | 2022-03-15 | 盛视科技股份有限公司 | Hair segmentation method and system based on convolutional neural network |
| CN114444679A (en) * | 2020-11-06 | 2022-05-06 | 山东产研鲲云人工智能研究院有限公司 | Quantization method and system for binary input model, and computer-readable storage medium |
| CN114663313A (en) * | 2022-03-25 | 2022-06-24 | 清华大学 | Vision method, apparatus, device and medium for enhancing facial mottled features |
| CN115485767A (en) * | 2020-05-07 | 2022-12-16 | 谷歌有限责任公司 | Emit word timing using an end-to-end model |
| CN115545172A (en) * | 2022-11-29 | 2022-12-30 | 支付宝(杭州)信息技术有限公司 | Method and device for training neural network of graph with privacy protection and fairness taken into account |
| CN117058739A (en) * | 2023-08-28 | 2023-11-14 | 联通在线信息科技有限公司 | A face clustering update method and device |
| CN117054396A (en) * | 2023-10-11 | 2023-11-14 | 天津大学 | Raman spectrum detection method and device based on double-path multiplicative neural network |
| CN117173854A (en) * | 2023-09-13 | 2023-12-05 | 西安博深安全科技股份有限公司 | Coal mine open fire early warning method and system based on deep learning |
| CN120496152A (en) * | 2025-07-15 | 2025-08-15 | 四川交通职业技术学院 | Scenic spot tourist density statistics method |
Families Citing this family (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109271891A (en) * | 2018-08-30 | 2019-01-25 | 成都考拉悠然科技有限公司 | A kind of dynamic face supervision method and system |
| CN109284700B (en) * | 2018-09-05 | 2021-03-16 | 武汉斗鱼网络科技有限公司 | Method, storage medium, device and system for detecting multiple faces in image |
| CN109497986A (en) * | 2018-11-22 | 2019-03-22 | 杭州脉流科技有限公司 | Electrocardiogram intelligent analysis method, device, computer equipment and system based on deep neural network |
| CN110163169A (en) * | 2019-05-27 | 2019-08-23 | 北京达佳互联信息技术有限公司 | Face identification method, device, electronic equipment and storage medium |
| CN110717377B (en) * | 2019-08-26 | 2021-01-12 | 平安科技(深圳)有限公司 | Face driving risk prediction model training and prediction method thereof and related equipment |
| CN111401430B (en) * | 2020-03-12 | 2022-04-01 | 四川大学 | Image online classification method based on dual-channel deep neural network |
| CN112529040A (en) * | 2020-11-16 | 2021-03-19 | 深圳壹账通智能科技有限公司 | Model generation method and device, electronic equipment and medium |
| CN113487698B (en) * | 2021-07-27 | 2023-12-12 | 深圳赛安特技术服务有限公司 | Form generation method and device based on two-channel neural network model |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105426963A (en) * | 2015-12-01 | 2016-03-23 | 北京天诚盛业科技有限公司 | Convolutional neural network Training method and apparatus for human face identification and application |
| CN105426860A (en) * | 2015-12-01 | 2016-03-23 | 北京天诚盛业科技有限公司 | Human face identification method and apparatus |
| CN106682616A (en) * | 2016-12-28 | 2017-05-17 | 南京邮电大学 | Newborn-painful-expression recognition method based on dual-channel-characteristic deep learning |
| CN107577987A (en) * | 2017-08-01 | 2018-01-12 | 广州广电卓识智能科技有限公司 | Method, system and device for identity authentication |
| CN107871105A (en) * | 2016-09-26 | 2018-04-03 | 北京眼神科技有限公司 | Face authentication method and device |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105960657B (en) * | 2014-06-17 | 2019-08-30 | 北京旷视科技有限公司 | Facial Super-Resolution Using Convolutional Neural Networks |
| WO2017214968A1 (en) * | 2016-06-17 | 2017-12-21 | Nokia Technologies Oy | Method and apparatus for convolutional neural networks |
| CN106778684A (en) * | 2017-01-12 | 2017-05-31 | 易视腾科技股份有限公司 | deep neural network training method and face identification method |
-
2018
- 2018-04-04 CN CN201810299180.6A patent/CN108446666A/en active Pending
- 2018-08-13 WO PCT/CN2018/100152 patent/WO2019192121A1/en not_active Ceased
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105426963A (en) * | 2015-12-01 | 2016-03-23 | 北京天诚盛业科技有限公司 | Convolutional neural network Training method and apparatus for human face identification and application |
| CN105426860A (en) * | 2015-12-01 | 2016-03-23 | 北京天诚盛业科技有限公司 | Human face identification method and apparatus |
| CN107871105A (en) * | 2016-09-26 | 2018-04-03 | 北京眼神科技有限公司 | Face authentication method and device |
| CN106682616A (en) * | 2016-12-28 | 2017-05-17 | 南京邮电大学 | Newborn-painful-expression recognition method based on dual-channel-characteristic deep learning |
| CN107577987A (en) * | 2017-08-01 | 2018-01-12 | 广州广电卓识智能科技有限公司 | Method, system and device for identity authentication |
Cited By (34)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111027421A (en) * | 2019-11-26 | 2020-04-17 | 西安宏规电子科技有限公司 | Graph-based direct-push type semi-supervised pedestrian re-identification method |
| CN110992329A (en) * | 2019-11-28 | 2020-04-10 | 上海微创医疗器械(集团)有限公司 | Product surface defect detection method, electronic device and readable storage medium |
| CN110992329B (en) * | 2019-11-28 | 2023-06-30 | 上海微创卜算子医疗科技有限公司 | A product surface defect detection method, electronic equipment and readable storage medium |
| CN112883213B (en) * | 2019-11-29 | 2023-11-17 | 深圳云天励飞技术有限公司 | Picture archiving method and device and electronic equipment |
| CN112883213A (en) * | 2019-11-29 | 2021-06-01 | 深圳云天励飞技术有限公司 | Picture archiving method and device and electronic equipment |
| CN111198963A (en) * | 2019-12-11 | 2020-05-26 | 智慧眼科技股份有限公司 | Target retrieval method and device based on average characteristics and related equipment thereof |
| CN111144566B (en) * | 2019-12-30 | 2024-03-22 | 深圳云天励飞技术有限公司 | Training method for neural network weight parameters, feature classification method and corresponding device |
| CN111144566A (en) * | 2019-12-30 | 2020-05-12 | 深圳云天励飞技术有限公司 | Neural network weight parameter training method, feature classification method and corresponding device |
| CN111144332A (en) * | 2019-12-30 | 2020-05-12 | 深圳云天励飞技术有限公司 | A kind of picture gathering method, device and electronic device |
| CN111209839A (en) * | 2019-12-31 | 2020-05-29 | 上海涛润医疗科技有限公司 | Face recognition method |
| CN111209839B (en) * | 2019-12-31 | 2023-05-23 | 上海涛润医疗科技有限公司 | Face recognition method |
| CN111324589A (en) * | 2020-02-14 | 2020-06-23 | 中国平安人寿保险股份有限公司 | Log sending method and device based on similarity degree value and computer equipment |
| CN111582358A (en) * | 2020-04-30 | 2020-08-25 | 上海添玑网络服务有限公司 | Training method and device for house type recognition model and house type weight judging method and device |
| CN111582358B (en) * | 2020-04-30 | 2024-05-10 | 上海添玑网络服务有限公司 | Training method and device for house type recognition model, and house type weight judging method and device |
| CN115485767A (en) * | 2020-05-07 | 2022-12-16 | 谷歌有限责任公司 | Emit word timing using an end-to-end model |
| CN111612785B (en) * | 2020-06-03 | 2024-02-02 | 浙江大华技术股份有限公司 | Face picture quality assessment method, device and storage medium |
| CN111612785A (en) * | 2020-06-03 | 2020-09-01 | 浙江大华技术股份有限公司 | Method and device for evaluating quality of face picture and storage medium |
| CN111709890B (en) * | 2020-06-12 | 2023-11-24 | 北京小米松果电子有限公司 | Training method and device for image enhancement model and storage medium |
| CN111709890A (en) * | 2020-06-12 | 2020-09-25 | 北京小米松果电子有限公司 | An image enhancement model training method, device and storage medium |
| CN114004779A (en) * | 2020-07-27 | 2022-02-01 | 中移物联网有限公司 | Face quality evaluation method and device based on deep learning |
| CN114444679A (en) * | 2020-11-06 | 2022-05-06 | 山东产研鲲云人工智能研究院有限公司 | Quantization method and system for binary input model, and computer-readable storage medium |
| CN113657289A (en) * | 2021-08-19 | 2021-11-16 | 北京百度网讯科技有限公司 | Training method, device and electronic device for threshold estimation model |
| CN113657289B (en) * | 2021-08-19 | 2023-08-08 | 北京百度网讯科技有限公司 | Training method, device and electronic equipment for threshold estimation model |
| CN114187309A (en) * | 2022-01-11 | 2022-03-15 | 盛视科技股份有限公司 | Hair segmentation method and system based on convolutional neural network |
| CN114663313A (en) * | 2022-03-25 | 2022-06-24 | 清华大学 | Vision method, apparatus, device and medium for enhancing facial mottled features |
| CN115545172B (en) * | 2022-11-29 | 2023-02-07 | 支付宝(杭州)信息技术有限公司 | Method and device for training neural network of graph with privacy protection and fairness taken into account |
| CN115545172A (en) * | 2022-11-29 | 2022-12-30 | 支付宝(杭州)信息技术有限公司 | Method and device for training neural network of graph with privacy protection and fairness taken into account |
| CN117058739A (en) * | 2023-08-28 | 2023-11-14 | 联通在线信息科技有限公司 | A face clustering update method and device |
| CN117173854B (en) * | 2023-09-13 | 2024-04-05 | 西安博深安全科技股份有限公司 | Coal mine open fire early warning method and system based on deep learning |
| CN117173854A (en) * | 2023-09-13 | 2023-12-05 | 西安博深安全科技股份有限公司 | Coal mine open fire early warning method and system based on deep learning |
| CN117054396A (en) * | 2023-10-11 | 2023-11-14 | 天津大学 | Raman spectrum detection method and device based on double-path multiplicative neural network |
| CN117054396B (en) * | 2023-10-11 | 2024-01-05 | 天津大学 | Raman spectrum detection method and device based on double-path multiplicative neural network |
| CN120496152A (en) * | 2025-07-15 | 2025-08-15 | 四川交通职业技术学院 | Scenic spot tourist density statistics method |
| CN120496152B (en) * | 2025-07-15 | 2025-09-30 | 四川交通职业技术学院 | A statistical method for tourist density in scenic spots |
Also Published As
| Publication number | Publication date |
|---|---|
| CN108446666A (en) | 2018-08-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| WO2019192121A1 (en) | Dual-channel neural network model training and human face comparison method, and terminal and medium | |
| CN112200136B (en) | Method, device, computer-readable medium, and electronic device for identifying authenticity of certificate | |
| US20220327189A1 (en) | Personalized biometric anti-spoofing protection using machine learning and enrollment data | |
| CN111091075B (en) | Face recognition method, device, electronic device and storage medium | |
| CN111783629B (en) | Human face in-vivo detection method and device for resisting sample attack | |
| JP2022133378A (en) | Face biometric detection method, device, electronic device, and storage medium | |
| JP7191061B2 (en) | Liveness inspection method and apparatus | |
| US20240013572A1 (en) | Method for face detection, terminal device and non-transitory computer-readable storage medium | |
| WO2020244071A1 (en) | Neural network-based gesture recognition method and apparatus, storage medium, and device | |
| CN112232323A (en) | Face verification method and device, computer equipment and storage medium | |
| TW202105329A (en) | Face verification method and apparatus, server and readable storage medium | |
| CN108764041A (en) | The face identification method of facial image is blocked for lower part | |
| CN110378328B (en) | certificate image processing method and device | |
| CN113642639B (en) | Living body detection method, living body detection device, living body detection equipment and storage medium | |
| US9525866B1 (en) | Building a digital camera fingerprint from cropped or corrupted images | |
| CN112232324A (en) | Face fake verification method, device, computer equipment and storage medium | |
| CN112883959B (en) | Identity card integrity detection method, device, equipment and storage medium | |
| US20240127631A1 (en) | Liveness detection method and apparatus, and computer device | |
| US20240193987A1 (en) | Face liveness detection method, terminal device and non-transitory computer-readable storage medium | |
| WO2022126464A1 (en) | Face recognition method, robot, and storage medium | |
| CN115205943A (en) | Image processing method, image processing device, electronic equipment and storage medium | |
| KR20200083188A (en) | Method and apparatus for detecting liveness and object recognition method using same | |
| CN113221842A (en) | Model training method, image recognition method, device, equipment and medium | |
| TWI796063B (en) | Liveness detection method and system threrof | |
| CN110738174B (en) | Finger vein recognition method, device, equipment and storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18913215 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 18913215 Country of ref document: EP Kind code of ref document: A1 |