WO2010126120A1 - Dispositif de traitement d'image, procédé de traitement d'image et programme destiné à amener un ordinateur à exécuter le programme - Google Patents
Dispositif de traitement d'image, procédé de traitement d'image et programme destiné à amener un ordinateur à exécuter le programme Download PDFInfo
- Publication number
- WO2010126120A1 WO2010126120A1 PCT/JP2010/057640 JP2010057640W WO2010126120A1 WO 2010126120 A1 WO2010126120 A1 WO 2010126120A1 JP 2010057640 W JP2010057640 W JP 2010057640W WO 2010126120 A1 WO2010126120 A1 WO 2010126120A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- target image
- concealment
- similarity
- sample point
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
Definitions
- the present invention relates to an image processing apparatus, an image processing method, and a program that causes a computer to execute the method, and in particular, regardless of the type of concealment that conceals an object in an input target image obtained by imaging the object.
- the present invention relates to an image processing apparatus capable of detecting the presence or absence of a concealment, an image processing method, and a program for causing a computer to execute the method.
- a face authentication device an authentication device that extracts a face image including a person's face from an input image obtained by capturing an image of the subject person and compares the extracted face image with a previously registered face image. It has been known.
- a method for collating face images is to set a plurality of sample points for the input face image and the registered face image, adjust the position of the sample points in consideration of the correspondence between both images,
- a method is known in which similarity is calculated for each sample point based on local features such as the distribution of surrounding gray values and the distance between sample points, and the similarity of the entire face is calculated based on this. (For example, refer nonpatent literature 1, patent documents 1 and 2).
- the sample point setting method is based on the sample points set on the reference average face image, temporarily setting the sample points on the input face image and the registered face image, and changing the position of the sample points, A method is known in which the position of the sample point corresponding to the sample point of the average face image is searched and determined (see Patent Documents 1 and 2).
- such an authentication apparatus compares the feature quantities of each part of the characteristic parts of the face extracted from both the input image and the registered image, and authenticates the person based on the similarity.
- a wearing object such as a mask or sunglasses is attached to the subject's face
- the authentication device cannot compare the part, so even the person himself / herself determines that it is another person. This is because the authentication accuracy is lowered.
- a general pattern recognition technique such as boosting (BOOSTING) or SVM (Super Vector Machine) is used to determine the presence or absence of an attachment.
- SVM Super Vector Machine
- a technique has been devised that determines the presence or absence of an attachment depending on whether a specific organ of the face such as the eyes or mouth can be detected (see, for example, Patent Document 4).
- the present invention has been made in view of the above, and provides an image processing apparatus, an image processing method, and a program for causing a computer to execute the method, capable of determining the presence or absence of a concealment object regardless of the type of the concealment object. For the purpose.
- the input is performed by comparing local features between a reference target image related to the target object and an input target image obtained by capturing the target object.
- An image processing method for detecting a concealment for concealing an object in a target image wherein a reference target image sample point setting step for setting a plurality of sample points on the reference target image, and setting on the reference target image A sample point corresponding to the sampled point is set on the input target image, a feature amount at each sample point on the reference target image and a feature amount at each sample point on the input target image are detected, and A similarity calculation step of calculating a similarity between the feature quantity of the reference target image and the feature quantity of the input target image for each sample point to be sampled, and a sample whose similarity is less than a predetermined similarity determination threshold Based on Le point, it was decided to provide an image processing method and a concealer presence determination step of determining whether the concealed object in the input object image. Thereby, the presence of the concealed object in the input object image.
- the sample in which all the sample points or in a group of the sample points selected in advance is less than the similarity determination threshold value.
- the number of points is larger than a predetermined concealment presence / absence determination threshold, it is determined that the concealment is present. Thereby, the presence or absence of the concealment can be detected regardless of the type of the concealment.
- the group is associated with a specific concealment, and when it is determined that the concealment is present in the group, it is determined that the associated specific concealment has been detected. Thereby, the presence or absence of the concealment can be detected regardless of the type of the concealment, and the type of the concealment can be estimated.
- a feature vector at the sample point is used as the feature amount, and the similarity of the feature vector is calculated as the similarity.
- the image processing method further includes a partial space creating step of creating a partial space from feature vectors at sample points set on a plurality of the reference target images, and the similarity calculation step includes: The partial space at the sample point is used as the feature quantity of the reference target image, the feature vector at the sample point is used as the feature quantity of the input target image, and the similarity between the partial space and the feature vector is used as the similarity. Calculate the degree.
- a concealment that conceals an object in the input target image by comparing local characteristics between a reference target image related to the object and an input target image obtained by capturing the target object.
- a program for causing a computer to execute an image processing method to be detected wherein sample points corresponding to a plurality of sample points set on the reference target image are set on the input target image, and The feature amount at the sample point and the feature amount at each sample point on the input target image are detected, and the similarity between the feature amount of the reference target image and the feature amount of the input target image is calculated for each corresponding sample point.
- a concealment that conceals an object in the input target image by comparing local characteristics between a reference target image related to the object and an input target image obtained by capturing the target object.
- An image processing apparatus for detecting wherein sample points corresponding to a plurality of sample points set on the reference target image are set on the input target image, and a feature amount at each sample point on the reference target image
- a similarity calculation unit that detects a feature amount at each sample point on the input target image and calculates a similarity between the feature amount of the reference target image and the feature amount of the input target image for each corresponding sample point;
- a concealment presence / absence determination unit that determines the presence / absence of the concealment in the input target image based on a sample point whose similarity is less than a predetermined similarity determination threshold is provided.
- the present invention it is possible to detect the presence or absence of a concealment object regardless of the type of the concealment object.
- FIG. 1 is an explanatory diagram showing an outline of a processing procedure for determining the presence or absence of sunglasses.
- FIG. 2 is an explanatory diagram showing an outline of a processing procedure for determining the presence or absence of a mask.
- FIG. 3 is an explanatory diagram showing an outline of a processing procedure when there is no concealment.
- FIG. 4 is an explanatory diagram for explaining the concept of the image recognition processing according to the present embodiment.
- FIG. 5 is an explanatory diagram for explaining the concept of the partial space according to the present embodiment.
- FIG. 6 is an explanatory diagram for explaining the concept of sample point detection on a face image using the distance of the feature vector to the recognition subspace according to the present embodiment.
- FIG. 1 is an explanatory diagram showing an outline of a processing procedure for determining the presence or absence of sunglasses.
- FIG. 2 is an explanatory diagram showing an outline of a processing procedure for determining the presence or absence of a mask.
- FIG. 3 is an explanatory diagram showing an outline of a
- FIG. 7 is an explanatory diagram for explaining the concept of the correlation value calculation processing of the feature vector on the average face image and the input image according to the present embodiment.
- FIG. 8 is a functional block diagram showing the configuration of the image recognition apparatus according to this embodiment.
- FIG. 9 is a flowchart of normalization processing for normalizing face image data according to the present embodiment.
- FIG. 10 is a flowchart of a recognition subspace creation process for creating a feature vector recognition subspace from a face image set according to the present embodiment.
- FIG. 11 is a flowchart of the recognition subspace correction process for correcting the recognition subspace of the feature vector from the face image set according to the present embodiment.
- FIG. 12 is a flowchart illustrating a processing procedure of the collation determination processing according to the present embodiment.
- FIG. 13 is a block diagram showing a system configuration of the image recognition apparatus according to the present embodiment.
- an embodiment of an image processing apparatus, an image processing method, and a program for causing a computer to execute the method according to the present invention will be specifically described with reference to the drawings.
- an example in which the present invention is applied to an image recognition apparatus that determines the presence or absence of a concealment that prevents face recognition in an input image obtained by capturing a human face as an object will be described. .
- the image recognition apparatus is configured to provide a concealment that conceals an object in the input target image by comparing local features between a reference target image related to the object and an input target image obtained by capturing the object. To detect.
- the imaging target for determining the presence / absence of the concealment is not limited to the person's face, but may be another part of the person such as a fingerprint or a palm print, or an arbitrary object.
- sunglasses and a mask are assumed as an example of a concealment, but the present invention assumes any concealment that may conceal a part of the imaging target, such as bangs or hands. Is.
- FIG. 1 is an explanatory diagram illustrating an outline of a processing procedure when determining the presence or absence of sunglasses
- FIG. 2 is an explanatory diagram illustrating an overview of a processing procedure for determining the presence or absence of a mask
- FIG. 4 is an explanatory diagram showing an outline of a processing procedure when there is no image
- FIG. 4 is an explanatory diagram for explaining a concept of image recognition processing according to the present embodiment.
- the image recognition apparatus when the image recognition apparatus according to the present embodiment receives an input image 16 in which an imaging target wearing sunglasses G is captured, first, the input target image (hereinafter referred to as “input face” in the input image 16 is received.
- the input image 16 shown in FIG. 1A is generated by normalizing the size of “image” and matching the size of the reference target image prepared in advance.
- the reference target image is an image obtained by capturing an object on which no concealment object is mounted.
- an average face image is used as the reference target image.
- the average face image is generated by normalizing a plurality of face images in the image set and averaging the gray value of each pixel.
- the reference target image is not limited to the averaged image as described above, but an averaged image is desirable when handling different targets such as a face.
- the average face image since common features of many faces are emphasized, it is possible to determine the likelihood of a face while suppressing the influence of individual features on various face images.
- the person's face image may be used.
- the image recognition apparatus sets 100 sample points in a region including both eyes, nose, and mouth in the average face image as shown in 12 of FIG. 4, and as shown in FIG. 100 sample points are also set in an area including both eyes, nose and mouth of the input face image.
- the image recognition apparatus is configured so that the position of each sample point set in the average face image and the position of each sample point set in the input face image are the positions of the corresponding portions of both face images. Set the position of the sample point.
- the image recognition device calculates the feature amount of each sample point in the corresponding portion of the average face image and the input face image, and compares the feature amounts of each sample point in the corresponding portion, thereby The similarity of the feature quantity is calculated.
- the similarity indicates how similar the average face image and the input face image are at the position of the sample point. Note that the concept of the feature amount regarding each sample point and the calculation method thereof will be described in detail later.
- FIG. 1C visually shows the degree of similarity with the average face image at each sample point set in the input face image. In the figure, the darker the color of the sample point, the higher the degree of similarity with the average face image at that sample point.
- the similarity of the feature amount of the region corresponding to the mouth and nose of the input face image is relatively high, and the similarity of the feature amount of the region corresponding to both eyes of the input face image. Is relatively low.
- the image recognition device determines the presence / absence of the concealment based on the numerical value of the similarity at each sample point set in the input face image.
- FIG. 3D shows the similarity value at each sample point on the input face image corresponding to FIG.
- the image recognition apparatus calculates the similarity at each sample point, and detects a sample point having a similarity less than a predetermined threshold (here, 40) from the sample points. .
- this threshold corresponds to the predetermined similarity determination threshold in the present invention.
- the image recognition apparatus detects the presence of the concealment object based on the ratio of the concealment area A1 in which the sample points having the similarity less than the threshold are present in the input face image. For example, the image recognition apparatus determines that there is a concealment object when the number of sample points constituting the concealment area A1 exceeds 30% of the entire input face image.
- the image recognition apparatus can estimate which part of the face is covered by the existing concealment object by determining the shape and location of the concealment area A1.
- the image recognition apparatus estimates that both eyes are concealed. To do.
- sample points can be selected according to the type of concealment to be detected, and the presence or absence of concealment can be determined based on the number of sample points that fall below the threshold in the selected group of sample points. For example, when detecting the sunglasses G, the sample point group B1 corresponding to both eyes is selected, and when the number of sample points below the threshold is equal to or greater than a predetermined number (or ratio), it is determined that the sunglasses G is worn.
- the image recognition apparatus may be configured to determine that there is a concealment object when the concealment presence / absence determination threshold value is greater.
- a predetermined group of sample points is associated with a specific concealment, and when it is determined that there is a concealment in the group, the associated specific concealment is
- the image recognition apparatus may be configured to determine that it has been detected.
- FIG. 2 shows a case where the imaging target is wearing the mask M.
- FIG. 2A shows the normalized input image 16 and
- FIG. 2B shows the input image 16 in which sample points are set.
- the image recognition apparatus calculates the similarity at each sample point, and the sample point having a similarity less than a predetermined threshold (here, 40) is selected from the sample points. And a concealment area A2 composed of these sample points is detected.
- a predetermined threshold here, 40
- the image recognition apparatus determines that there is a concealment object because the number of sample points constituting the concealment area A2 exceeds 30% of the entire input face image. Further, since the concealment area A2 that conceals 80% or more of the sample point group B2 corresponding to the mouth is detected, the image processing apparatus estimates that the mouth is hidden.
- the group C2 of sample points corresponding to the region including the mouth and nose is selected, and the number of sample points that fall below the threshold among the sample points in the group C2 is equal to or greater than a predetermined number (or ratio). At this time, it may be determined that the mask M is worn.
- FIG. 3 shows a case where the imaging target is not wearing anything on the face.
- FIG. 3A shows the normalized input image 16 and
- FIG. 3B shows the input image 16 in which sample points are set.
- the image recognition apparatus calculates the similarity at each sample point, and the sample point from which the similarity is less than a predetermined threshold value (here, 40) is calculated. Is detected, and a concealed area A3 composed of these sample points is detected.
- a predetermined threshold value here, 40
- the image recognition apparatus determines that the concealment area A3 is not a concealment object because the number of sample points constituting each concealment area A3 does not reach 30% of the entire input face image.
- a threshold value may be set separately for the group B3 of sample points corresponding to the cheeks. That is, the threshold value may be set to a different value for each sample point.
- the image recognition apparatus sets a plurality of sample points at corresponding portions of the input face image and the average face image, and compares image feature amounts at the corresponding sample points. The similarity is calculated.
- the image recognition apparatus estimates that the area is hidden by the concealment object when the similarity related to the predetermined area in the input face image is lower than a certain threshold.
- this image recognition apparatus does not need to detect the concealment by performing pattern matching using various concealment image patterns assumed in advance on the input face image in order to determine the presence or absence of the concealment. Therefore, the processing amount required for pattern matching can be reduced.
- this image recognition apparatus does not detect a concealment, but the similarity between the sample point group set in the normalized input image and the average face image at the sample point falls below a predetermined threshold. In this case, since it is determined that the area is hidden by the concealment object, the presence / absence of the concealment object can be detected even if the concealment object is not assumed in advance.
- the image recognition apparatus can determine the presence of the concealment without detecting the concealment as described above, for example, when the eyes are concealed by long bangs, Even if a part of the cover is concealed, the bangs and hands can be determined as a concealment, so the presence or absence of the concealment can be determined regardless of the type of concealment.
- the type of concealment can be estimated based on the positions and shapes of the concealment areas A1 and A2 configured by sample points whose feature amount similarity is less than a predetermined threshold.
- the image recognition device when the image recognition device is applied to a personal authentication device using a face image, the user is pointed out a concealment (for example, sunglasses G, mask M, etc.) worn on the face, Care can be taken to remove the cover.
- a concealment for example, sunglasses G, mask M, etc.
- FIG. 4 is an explanatory diagram for explaining the concept of the image recognition processing according to the present embodiment.
- the image recognition process for determining the presence / absence of the concealment in the input face image an image recognition process for determining the presence / absence of the concealment in the face authentication process will be described as an example.
- the similarity (local similarity) at the sample points described above indicates how similar the local images near the corresponding sample points of the two images are. Generally, it means that it is similar if the similarity is large.
- a method is used in which the local images are overlapped to obtain the degree of coincidence, or the feature amounts of the local images are vectorized and compared.
- the method of comparing feature vectors includes a method of comparing feature vectors and a method of comparing a feature space with a partial space created from a plurality of feature vectors.
- the distance and angle between feature vectors are obtained, and the local similarity is calculated. If this distance or angle is small, it means that the feature vectors are similar.
- the similarity is first obtained for each image, and the local similarity is obtained by averaging the obtained similarities.
- there are multiple reference images a subspace is created from a plurality of feature vectors obtained from these, the distance and angle between this subspace and the feature vector, the length of the feature vector projected onto the subspace, etc. And the local similarity is calculated.
- the distance or angle is small or the length is long, it means that the feature vector is similar to the subspace.
- the Euclidean distance, the Maranobis distance, etc. can be used as the distance used when comparing the feature vectors.
- any method other than those described above can be used to calculate the local similarity.
- the presence or absence of a concealment on the face image of the input image is determined using the local similarity obtained in the face authentication process.
- the presence / absence of a concealment is determined using the local similarity obtained based on the angle and distance of the feature vector and the distance to the partial space.
- the reference target image an average face image, a plurality of face images, and a registered image for comparison are used.
- the image recognition process in the present embodiment is composed of two major processes, an offline learning process and a collation process, and uses knowledge 13 obtained by learning the features of the face image in the learning process.
- the process for determining the presence or absence of the concealment is performed during the learning process or the collation process.
- the processing procedure for offline learning will be described.
- the positions of both eyes, nose, and mouth are designated for each image in the face image set 10.
- the positions of both eyes should be specified at least here.
- normalization is performed so that the positions of both eyes, nose, and mouth of each face image overlap, and the average face image 11 is created by averaging the gray value of each pixel.
- the “image set” is a collection of images related to a predetermined part of the body of a plurality of persons, and a plurality of different persons also differ depending on differences in facial expressions, palm open / close, imaging angle, etc.
- the person's own image is included.
- FIG. 4 shows an example 12 of the average face image 11 in which 100 sample points are designated on the average face image 11. Then, using the correlation between the feature vector of the average face image 11 and the feature vector of each image in the face image set 10, sample points on each image in the face image set 10 corresponding to the sample points on the average face image 11 are used. Is detected.
- a sample point is set at a position on each image in the face image set 10 corresponding to the sample point on the average face image 11, and the correlation between the feature vectors is obtained. Then, the position where the correlation value increases is obtained while shifting the sample point, and the position of the sample point is determined.
- the correlation value of the feature vector obtained here corresponds to the local similarity at the sample point of each image obtained using the average face image 11 as the reference target image. Specifically, the value corresponds to the similarity shown in (d) of FIGS. By using this, it is possible to detect the presence or absence of a concealment on each image in the face image set 10 by the method described above.
- a partial space of the feature vector in each image in the face image set 10 is created using a principal component analysis technique. This subspace is provisional and will be modified later.
- the “feature vector” is a collection of feature quantities such as a gray value of a pixel in a local region around the sample point and a change rate of the gray value as a vector composed of real elements. If a digital face image composed of a collection of a plurality of pixels each having these feature amounts is given, it can be easily created.
- FIG. 5 is an explanatory diagram for explaining the concept of the partial space according to the present embodiment.
- the person's variation subspace 22 is composed of differences in feature vectors of a plurality of person image pairs that differ depending on differences in facial expressions, photographing angles, and the like.
- the other person's variable subspace 21 is composed of the difference in the feature vector of the image pair between the person and the other person.
- the target subspace 23 can be derived from the intersection of the subspace of the person's variable subspace 22 and the other person's variable subspace 21.
- the partial space configured in this way is referred to as a “recognition partial space”.
- the provisional recognition subspace 23 of the feature vector created using the images in the face image set 10 is transferred to the feature vector recognition subspace 23 in each image in the face image set 10. It is corrected based on the distance.
- the distance to the recognition subspace 23 of the feature vector in each image in the face image set 10 is calculated, and a sample point on each image that decreases the distance is detected.
- a new recognition subspace 23 is created based on the feature vector in each image of the sample point.
- the partial space creation process is repeated, and the recognition partial space 23 is corrected.
- the obtained recognition subspace 23 is used at the time of image collation as knowledge 13 obtained by learning the features of the face image.
- FIG. 6 is an explanatory diagram for explaining the concept of sample point detection on a face image using the distance of the feature vector to the recognition subspace 23 according to the present embodiment.
- sample points corresponding to the sample points on the average face image 11 can be set on the image. That is, by reducing this distance, the consistency of the position of the input face image and the average face image 11 is improved when determining the concealment.
- FIG. 4 shows an example 15 of a registered face image 14 in which sample points are detected on the registered face image 14.
- the input image 16 is also calculated by calculating the distance of the feature vector in the input image 16 to the recognition subspace 23 and detecting a sample point on the input image 16 where the distance becomes small. To do.
- FIG. 4 shows an example 17 of the input image 16 in which sample points are detected on the input image 16.
- the distance to the recognition subspace 23 obtained here corresponds to the local similarity at the sample points obtained from the plurality of face images used for creating the recognition subspace 23 as reference target images. Specifically, the value corresponds to the similarity shown in (d) of FIGS. Using this, it is possible to detect the presence or absence of a concealment on the input image 16 by the method described above.
- the feature vector of the registered face image 14 and the feature vector of the input image 16 projected into the recognition subspace 23 at the sample points are used next.
- the correlation value of both feature vectors is calculated in the recognition subspace 23. This process is performed for all sample points.
- the average value of the correlation values for the image pair of the input image 16 and the registered image 14 is calculated and set as the similarity of the whole image.
- Example 15 in which 100 sample points are set on the registered face image 14, an average value of 100 correlation values (similarity of the entire image) is calculated. Then, based on the similarity of the whole image, the person and others are identified.
- FIG. 7 is an explanatory view for explaining the concept of the feature vector correlation value calculation processing on the registered face image 14 and the input image 16 according to the present embodiment. As shown in FIG. 7, the feature vector of the input image 16 and the feature vector of the registered face image 14 are projected onto the recognition subspace 23.
- the correlation value between the feature vector 40 of the projected input image 16 and the feature vector 41 of the projected registered face image 14 is calculated by the cosine of the angle ⁇ formed between the projected feature vectors, that is, cos ⁇ .
- a larger value means that the features of both images at the sample point are similar.
- the correlation value based on the angle ⁇ formed by the both projected feature vectors obtained here corresponds to the local similarity at the sample point obtained using the registered face image 14 as a reference target image. Specifically, the value corresponds to the similarity shown in (d) of FIGS. Again, using this, the presence or absence of a concealment on the input image 16 can be detected by the method described above.
- the projection onto the recognition subspace 23 is a process for identification determination. If only the concealment is detected, the correlation value of the unprojected feature vector may be obtained.
- FIG. 8 is a functional block diagram showing the configuration of the image recognition apparatus according to this embodiment.
- the image recognition apparatus 50 includes a face image input receiving unit 51, a normalization processing unit 52, a face image learning unit 53, a registered image storage unit 54, and a registered image.
- a sample point detection unit 55, an input image sample point detection unit 56, a similarity calculation unit 57, a collation determination unit 58, and a concealment determination unit 59 are provided.
- the face image input receiving unit 51 is a receiving unit for taking a learning face image set, an input face image, and a registration face image into the apparatus, and outputs the received image to the normalization processing unit 52.
- the normalization processing unit 52 acquires an image from the face image input receiving unit 51, normalizes the image to match the size of the face of each image, and the received image is an image of a learning face image set. If there is a registered face image, the normalized image is output to the registered image storage unit 54, and if it is an input face image, the normalized image is output to the input image sample point detecting unit 56.
- the face image learning unit 53 acquires a normalized face image set from the normalization processing unit 52, creates an average face image 11 using the face image set 10, and uses the average face image 11 to generate a feature vector.
- the recognition subspace 23 is created.
- the registration image storage unit 54 acquires and stores the registration image normalized by the normalization processing unit 52 when the registration face image is received by the face image input reception unit 51.
- the registration image sample point detection unit 55 reads the recognition subspace 23 from the face image learning unit 53 and reads the registration image from the registration image storage unit 54 when collation determination is performed between the registration image and the input image 16.
- the sample points on the registered image are detected using the distance of the feature vector in the registered image to the recognition subspace 23.
- the input image sample point detection unit 56 reads the recognition subspace 23 from the face image learning unit 53 and receives the input image 16 from the normalization processing unit 52 when collation determination is performed between the registered image and the input image 16.
- the sample point on the input image 16 is detected using the distance of the feature vector in the input image 16 to the recognition subspace 23.
- the similarity calculation unit 57 reads the sample point information of the registered image 14 and the input image 16 and the feature vector information of the registered image 14 and the input image 16 at the sample point, and registers the input image 14 and the input image at each corresponding sample point. The local similarity between 16 is calculated.
- the collation determination unit 58 reads the overall similarity between the registered image and the input image 16 obtained by the similarity calculation unit 57, performs collation determination between the registered image and the input image 16 based on the overall similarity, Output the verification result.
- the concealment determination unit 59 reads the local similarity between the recognition subspace 23 obtained by the input image sample point detection unit 56 and the input image 16, or the registration image 14 obtained by the similarity calculation unit 57 and the input. The local similarity with the image 16 is read, and based on the local similarity, it is determined whether or not a concealment exists on the input face image in the input image 16, and the determination result is output.
- FIG. 9 is a flowchart of normalization processing for normalizing face image data according to the present embodiment. This processing is performed by the normalization processing unit 52. Here, a case where the image received by the face image input receiving unit 51 is a learning face image set is shown.
- an average face image 11 is provisionally provided, and provisional positions of both eyes are set on the image (step S101).
- the face image data in the face image set 10 is read through the face image input receiving unit 51 (step S102), and the positions of both eyes, nose and mouth of each read face image data are designated (step S103).
- the positions of both eyes should be specified at least here.
- affine transformation is performed so that the positions of both eyes of each face image data overlap with the positions of both eyes of the average face image 11 (step S104).
- the normal positions of both eyes, nose and mouth are calculated (step S105).
- the average value of the calculated normal positions of both eyes, nose and mouth is calculated (step S106), and the positions of both eyes, nose and mouth of the average face image 11 are set (step S107).
- Step S108 affine transformation is performed so that the positions of both eyes, nose and mouth of each face image data overlap with the positions of both eyes, nose and mouth of the average face image 11.
- Step S109 a normalized image of each face image in the image set 10 is created.
- the process at the time of learning ends and the received image is the input face image. Since the positions of the eyes, nose, and mouth of the average face image 11 have already been set during learning, the face image data is read (step S102), and both eyes of the face image data read based on the learned knowledge, The positions of the nose and mouth are designated (step S103), and affine transformation is performed so that the positions of both eyes, nose and mouth of the face image data overlap with the positions of both eyes, nose and mouth of the average face image 11 (step S108), and read. It is only necessary to create a normalized image of the image (step S109).
- FIG. 10 is a flowchart of a recognition subspace creation process for creating a feature vector recognition subspace 23 from the face image set 10 according to the present embodiment.
- N face image data are read from the learning face image set 10 (step S201).
- the average of the shade values of the corresponding pixels of the read N face image data is calculated (step S202), and the average face image 11 is created (step S203).
- M sample points are set on the created average face image 11 (step S204).
- step S205 M sample points corresponding to the sample points on the average face image 11 are set on each face image data in the N face image sets 10 (step S205), and the corresponding M sample points are set.
- the correlation is calculated from the angle and distance between the average face image 11 and the feature vectors of the N face image data (step S206).
- step S207 it is checked whether or not the correlation between the feature vectors is larger than a certain threshold value. If the correlation is large, M sample points of the average face image 11 are added to the N face image data. Corresponding sample points are determined (step S208). If the correlation is small, M sample points are reset on each of the N face image data (step S205), and these processes (steps S205 to S207) are repeated until a sample point with a large correlation is found. . If the correlation does not become larger than the threshold value even after being repeated a predetermined number of times, the maximum value is set as the correlation value.
- the concealment of each face image can be detected.
- the feature vector correlation calculated in step S206 corresponds to the local similarity at the sample points of the average face image 11 and each face image data. Therefore, the presence or absence of the concealment of each face image can be detected by the method of the present invention.
- FIG. 11 is a flowchart of a recognition subspace correction process for correcting the feature vector recognition subspace 23 from the face image set 10 according to the present embodiment.
- step S301 the maximum number of iterations of the convergence calculation of the recognition subspace 23 is set (step S301), and M sample points on the average face image 11 are set on each of the N face images in the learning face image set 10. Corresponding M sample points are set (step S302). Then, the distances to the M recognition subspaces 23 corresponding to the M feature vectors in the N face images are calculated (step S303).
- step S304 it is checked whether or not the distance to the recognition subspace 23 is smaller than a certain threshold value (step S304). If the distance is small, M samples on the average face image 11 are included in each N face image data. A sample point corresponding to the point is determined (step S305). If the distance is large, M sample points are set again on each of the N face image data (step S302), and these processes (steps S302 to S304) are repeated until a sample point with a small distance is found. .
- step S305 When sample points corresponding to the M sample points on the average face image 11 are determined on each of the N face image data (step S305), the sample points at the M sample points on the N face image data are determined. M feature vector recognition subspaces 23 are created (step S306). Then, it is checked whether or not the recognition subspace 23 has converged (step S307). If the recognition subspace 23 has converged, the correction processing of the recognition subspace 23 is finished assuming that learning has ended.
- step S308 it is next checked whether or not the number of iterations of convergence calculation is smaller than the set maximum number of iterations. If not, the correction processing of the recognition subspace 23 is finished. In this case, M sample points are set again on each of the N face image data (step S302), and these processes (steps S302 to S308) are repeated until the recognition subspace 23 converges.
- FIG. 12 is a flowchart illustrating the processing procedure of the collation determination process according to the first embodiment.
- the registered image sample point detection unit 55 reads the registered image stored in the registered image storage unit 54 and the data of the recognition subspace 23 created by the face image learning unit 53, and reads the average face image 11 on the registered image. M sample points corresponding to the upper sample point are set (step S401). Then, the distance to the recognition subspace 23 of the feature vector of the registered image is calculated on each corresponding sample point (step S402). This concept of distance has already been explained in FIG.
- step S403 it is checked whether or not the distance to the recognition subspace 23 is smaller than a certain threshold value (step S403). If the distance is small, M sample points of the average face image 11 on each of the N face image data. A sample point corresponding to is determined (step S404). If the distance is not small, M sample points are set again on each of the N face image data (step S401), and these processes (steps S401 to S403) are performed until a sample point with a small distance is found. repeat. These processes (steps S401 to S404) are performed by the registered image sample point detection unit 55.
- the input image sample point detection unit 56 reads the input image 16 normalized by the normalization processing unit 52 and the recognition subspace data 23 created by the face image learning unit 53, and puts it on the input image 16. M sample points corresponding to the sample points on the registered face image 14 are set (step S405). Then, the distance of the feature vector of the input image 16 to the recognition subspace 23 is calculated on each corresponding sample point (step S406).
- step S407 it is checked whether or not the distance to the recognition subspace 23 is smaller than a certain threshold (step S407). If the distance is small, it corresponds to M sample points of the registered face image 14 on the face image data. A sample point is determined (step S408). If the distance is not small, M sample points are set again on the face image data (step S405), and these processes (steps S405 to S407) are repeated until a sample point with a small distance is found. These processes (steps S405 to S407) are performed by the input image sample point detector 56.
- the local image that does not have facial features is compared with the recognition subspace, so the distance obtained in S406 is smaller than the threshold value. It may not be possible. In this case, it is checked whether or not the obtained distance has converged to the minimum value, and if it has converged, the sample point is determined.
- the similarity calculation unit 57 inputs the correlation value data of the registered image and the input image 16 between the determined sample points as the input image sample. Reading from the point detection unit 56, the local similarity (correlation value in the recognition subspace) between the registered image and the input image 16 at each sample point is calculated (step S409).
- the concealment determination unit 59 determines whether or not there is a concealment on the input image based on the local similarity between the registered image and the input image 16 at each sample point calculated in step S409 ( Step S410).
- the collation determination unit 58 performs collation determination between the registered image and the input image 16 based on the magnitude of the local similarity calculated in step S409 (step S412). ), A determination result as to whether or not the person corresponds to the registered image is output, and the process is terminated.
- the concealed object determination unit 59 outputs a determination result to give a warning, and performs a collation determination between the registered image and the input image 16 (step S411), and ends the process. To do.
- sample points are determined while comparing the threshold value and the correlation (local similarity). However, a certain number of sample points or more are always set and the correlation is calculated. The point where the correlation is maximum may be used as the sample point. Thereby, although the processing amount may increase, the corresponding sample point and the local similarity at the sample point can be obtained more accurately.
- FIG. 13 is a block diagram showing the system configuration of the image recognition apparatus according to this embodiment.
- the image recognition device 50 includes an interface unit 64, a calculation control unit 65, a main storage unit 66, and an auxiliary storage unit 67, and an input device 68 and a display device 69 are connected to each other. This is a stand-alone configuration using a computer.
- a CCD camera 60 for capturing the input image 16 is connected to the image recognition device 50 via an interface unit 64.
- the CCD camera 60 includes an image input unit 61, an image memory 62, and an interface unit 63. It consists of
- the image input unit 61 collects light from the face as a subject with a lens, converts the face image into an electrical signal using a CCD (Charge Coupled Device), converts the face image into digital data, and converts the face image data into an image. Record in the memory 62.
- CCD Charge Coupled Device
- the image memory 62 is used as a buffer memory of the image input unit 61, and temporarily stores the face image data when the image recognition device 50 cannot accept the input of the face image.
- face image data is output to the image recognition device 50 through the interface unit 63.
- the interface unit 64 receives the input image 16 from the CCD camera 60, receives data from the input device 68, and transfers data to the display device 69 under the control of the arithmetic control unit 65.
- the input device 68 includes a keyboard, a mouse, a touch panel, and the like.
- the display device 69 is a display monitor, and is used for displaying sample points on the average face image 11, displaying sample points on the input image 16, or displaying a concealment determination result when executing the image recognition program.
- the auxiliary storage unit 67 can read data from, for example, a floppy (registered trademark), a disk drive device (FDD), a hard disk drive device (HDD), a CD-ROM, a CD-R, a CD-RW, or the like.
- a DVD drive device that can read data from a DVD-ROM, DVD-R, DVD-RW, DVD-RAM, or the like.
- the image recognition program executed by the image recognition device 50 is recorded and provided as a file in an executable format on an FD, CD-ROM, DVD-ROM, or the like.
- the program is read and executed by a floppy (registered trademark) disk drive device, a CD drive device, a DVD drive, or the like.
- the face image data in the learning face image set 10 and the face image data for registration are also provided by FD, CD-ROM, DVO-ROM or the like, and are a floppy (registered trademark) disk drive device, CD drive device or DVD.
- the face image data for learning read by the drive is stored and processed in the face image learning unit 53.
- auxiliary storage unit 67 does not need to be directly connected to the image recognition device 50, and the auxiliary storage unit 67 may be present on the network.
- a configuration may be adopted in which a face image server is installed on the Internet or a LAN, learning face image data or the like is stored, and downloaded as necessary.
- the image recognition device 50 needs to be further provided with a communication unit such as a modem or a LAN board.
- the image recognition apparatus 50 of this embodiment is provided with an arithmetic control unit 65 such as a CPU for controlling the entire system, and a main storage unit 66 composed of storage media such as a RAM and a ROM.
- a boot program and the like are stored in advance in the ROM, and a part of an OS (operation system) read from the HD, an image recognition program, and the like are stored in the RAM, and the arithmetic control unit 65 executes these programs.
- the RAM stores various face image data, sample point data, calculation results, and the like read when the image recognition program is executed.
- sample points are set at locations corresponding to the sample points set in the reference target image, and the reference at the sample points is set. By detecting a location where the local similarity with the target image is less than a predetermined threshold, it is determined that a concealment exists.
- the image recognition apparatus 50 can determine the presence / absence of the concealment without depending on the type of concealment in the input image 16, thereby reducing the processing load required to determine the presence / absence of the concealment. be able to.
- the image recognition device 50 can also estimate the type of the concealment based on the shapes of the concealment areas A1 to A3 detected from the input image 16.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Collating Specific Patterns (AREA)
Abstract
Selon l'invention, on peut détecter la présence/l'absence d'un objet dissimulant indépendamment du type d'objet dissimulant qui dissimule un sujet dans des images à entrer, le sujet ayant été capturé par image. On compare les caractéristiques locales entre une image à référencer associée au sujet et les images à entrer, le sujet ayant été capturé par image et, lorsqu'on détecte l'objet dissimulant qui dissimule le sujet dans les images à entrer, on établit une pluralité de points échantillons sur l'image à référencer, les points échantillons correspondant aux points échantillons établis sur l'image à référencer étant établis sur les images à entrer, on détecte la valeur caractéristique de chacun des points échantillons sur l'image à référencer et la valeur caractéristique de chacun des points échantillons sur les images à entrer, la similarité entre les quantités caractéristiques de l'image à référencer et les images à entrer étant calculée pour chacun des points échantillons correspondants, et la présence/absence de l'objet dissimulant dans les images à entrer étant déterminée sur la base des points échantillons pour lesquels la similarité est inférieure à une valeur seuil prédéterminée.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2009111315A JP5480532B2 (ja) | 2009-04-30 | 2009-04-30 | 画像処理装置、画像処理方法、及び同方法をコンピュータに実行させるプログラム |
| JP2009-111315 | 2009-04-30 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2010126120A1 true WO2010126120A1 (fr) | 2010-11-04 |
Family
ID=43032264
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2010/057640 Ceased WO2010126120A1 (fr) | 2009-04-30 | 2010-04-22 | Dispositif de traitement d'image, procédé de traitement d'image et programme destiné à amener un ordinateur à exécuter le programme |
Country Status (2)
| Country | Link |
|---|---|
| JP (1) | JP5480532B2 (fr) |
| WO (1) | WO2010126120A1 (fr) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPWO2017212967A1 (ja) * | 2016-06-08 | 2019-01-10 | パナソニックIpマネジメント株式会社 | 照合装置及び照合方法 |
| JP2021103538A (ja) * | 2019-12-12 | 2021-07-15 | 日本電気株式会社 | 情報処理装置、情報処理方法、および、情報処理プログラム |
| WO2022123751A1 (fr) | 2020-12-10 | 2022-06-16 | 富士通株式会社 | Procédé de détermination, programme de détermination et dispositif de traitement d'informations |
Families Citing this family (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP6431346B2 (ja) * | 2014-11-26 | 2018-11-28 | アルパイン株式会社 | 顔認識装置 |
| JP2021015317A (ja) * | 2017-11-21 | 2021-02-12 | 富士フイルム株式会社 | 認識装置、認識方法及びプログラム |
| JP7054847B2 (ja) * | 2019-03-04 | 2022-04-15 | パナソニックIpマネジメント株式会社 | 顔認証登録装置および顔認証登録方法 |
| CN110610127B (zh) * | 2019-08-01 | 2023-10-27 | 平安科技(深圳)有限公司 | 人脸识别方法、装置、存储介质及电子设备 |
| CN111026641B (zh) * | 2019-11-14 | 2023-06-20 | 北京云聚智慧科技有限公司 | 一种图片比较方法和电子设备 |
| CN112115886A (zh) * | 2020-09-22 | 2020-12-22 | 北京市商汤科技开发有限公司 | 图像检测方法和相关装置、设备、存储介质 |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2004221871A (ja) * | 2003-01-14 | 2004-08-05 | Auto Network Gijutsu Kenkyusho:Kk | 車輌周辺監視装置 |
| WO2005122093A1 (fr) * | 2004-06-07 | 2005-12-22 | Glory Ltd. | Dispositif de reconnaissance d’image, méthode de reconnaissance d’image et programme pour qu’un ordinateur applique la méthode |
Family Cites Families (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2005084815A (ja) * | 2003-09-05 | 2005-03-31 | Toshiba Corp | 顔認識装置、顔認識方法および通行制御装置 |
| JP4862447B2 (ja) * | 2006-03-23 | 2012-01-25 | 沖電気工業株式会社 | 顔認識システム |
-
2009
- 2009-04-30 JP JP2009111315A patent/JP5480532B2/ja not_active Expired - Fee Related
-
2010
- 2010-04-22 WO PCT/JP2010/057640 patent/WO2010126120A1/fr not_active Ceased
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2004221871A (ja) * | 2003-01-14 | 2004-08-05 | Auto Network Gijutsu Kenkyusho:Kk | 車輌周辺監視装置 |
| WO2005122093A1 (fr) * | 2004-06-07 | 2005-12-22 | Glory Ltd. | Dispositif de reconnaissance d’image, méthode de reconnaissance d’image et programme pour qu’un ordinateur applique la méthode |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPWO2017212967A1 (ja) * | 2016-06-08 | 2019-01-10 | パナソニックIpマネジメント株式会社 | 照合装置及び照合方法 |
| US11367308B2 (en) | 2016-06-08 | 2022-06-21 | Panasonic Intellectual Property Management Co., Ltd. | Comparison device and comparison method |
| JP2021103538A (ja) * | 2019-12-12 | 2021-07-15 | 日本電気株式会社 | 情報処理装置、情報処理方法、および、情報処理プログラム |
| JP7124912B2 (ja) | 2019-12-12 | 2022-08-24 | 日本電気株式会社 | 情報処理装置、情報処理方法、および、情報処理プログラム |
| WO2022123751A1 (fr) | 2020-12-10 | 2022-06-16 | 富士通株式会社 | Procédé de détermination, programme de détermination et dispositif de traitement d'informations |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2010262392A (ja) | 2010-11-18 |
| JP5480532B2 (ja) | 2014-04-23 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP5480532B2 (ja) | 画像処理装置、画像処理方法、及び同方法をコンピュータに実行させるプログラム | |
| KR102359558B1 (ko) | 얼굴 인증 방법 및 장치 | |
| KR102455633B1 (ko) | 라이브니스 검사 방법 및 장치 | |
| US11023757B2 (en) | Method and apparatus with liveness verification | |
| JP7004522B2 (ja) | ライブネス検査方法及び装置 | |
| JP2022071195A (ja) | コンピューティング装置及び方法 | |
| JP4156430B2 (ja) | データベース自動更新方法を用いた顔検証方法及びそのシステム | |
| JP2020074174A (ja) | モバイル・デバイスを用いてキャプチャしたイメージを使用する指紋ベースのユーザ認証を実行するためのシステムおよび方法 | |
| KR20190136349A (ko) | 3d 영상 기반의 얼굴 인증 방법 및 장치 | |
| KR20190001066A (ko) | 얼굴 인증 방법 및 장치 | |
| Pamplona Segundo et al. | Continuous 3D face authentication using RGB-D cameras | |
| CN110647955A (zh) | 身份验证方法 | |
| JP2020184331A (ja) | ライブネス検査方法及び装置、顔認証方法及び装置 | |
| WO2005122093A1 (fr) | Dispositif de reconnaissance d’image, méthode de reconnaissance d’image et programme pour qu’un ordinateur applique la méthode | |
| KR20190053602A (ko) | 얼굴 인증 방법 및 장치 | |
| JP6822482B2 (ja) | 視線推定装置、視線推定方法及びプログラム記録媒体 | |
| JP5170094B2 (ja) | なりすまし検知システム、なりすまし検知方法およびなりすまし検知用プログラム | |
| KR20190060671A (ko) | 홍채 인식 기반 사용자 인증 장치 및 방법 | |
| JP2004265267A (ja) | 顔認証方法、および顔認証装置。 | |
| JP5703068B2 (ja) | 顔検出方法および顔検出装置、ならびに、プログラム | |
| JP2012185769A (ja) | 認証装置、認証方法、および認証プログラム、並びに記録媒体 | |
| JP5748421B2 (ja) | 認証装置、認証方法、及び認証プログラム、並びに記録媒体 | |
| JP4187494B2 (ja) | 画像認識装置、画像認識方法およびその方法をコンピュータに実行させるプログラム | |
| KR102570071B1 (ko) | 라이브니스 검사 방법 및 장치, 얼굴 인증 방법 및 장치 | |
| KR20180108361A (ko) | 얼굴 인증 방법 및 장치 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10769818 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 10769818 Country of ref document: EP Kind code of ref document: A1 |