CN100361138C - Method and system for real-time detection and continuous tracking of human faces in video sequences - Google Patents
Method and system for real-time detection and continuous tracking of human faces in video sequences Download PDFInfo
- Publication number
- CN100361138C CN100361138C CNB2005101356688A CN200510135668A CN100361138C CN 100361138 C CN100361138 C CN 100361138C CN B2005101356688 A CNB2005101356688 A CN B2005101356688A CN 200510135668 A CN200510135668 A CN 200510135668A CN 100361138 C CN100361138 C CN 100361138C
- Authority
- CN
- China
- Prior art keywords
- face
- people
- detection
- image
- real
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Landscapes
- Image Analysis (AREA)
Abstract
本发明提出一种视频序列中人脸的实时检测与持续跟踪的方法及系统,该方法包括以下步骤:输入视频图像;采用实时人脸检测算法对输入的视频图像进行人脸检测;再采用粗细两级人脸检测算法对检测到的人脸进行检测验证;采用物体跟踪算法跟踪验证后的人脸;通过对跟踪区域的验证对跟踪的人脸进行验证处理。本发明采用基于AdaBoost统计分层分类器的人脸检测方法,实现正面直立人脸的实时检测,并结合基于Mean shift和直方图特征的人脸跟踪方法,实现了人脸的实时检测和跟踪,本发明具有检测及跟踪迅速、实时性强的优点。
The present invention proposes a method and system for real-time detection and continuous tracking of human faces in video sequences. The two-level face detection algorithm detects and verifies the detected faces; uses the object tracking algorithm to track the verified faces; and verifies the tracked faces through the verification of the tracking area. The present invention adopts the human face detection method based on the AdaBoost statistical layered classifier to realize the real-time detection of the frontal upright human face, and combines the human face tracking method based on Mean shift and histogram features to realize the real-time detection and tracking of the human face, The invention has the advantages of rapid detection and tracking and strong real-time performance.
Description
技术领域technical field
本发明涉及人脸检测和跟踪领域,尤其是指一种在视频序列中人脸的实时检测与持续跟踪的方法及系统。The invention relates to the field of human face detection and tracking, in particular to a method and system for real-time detection and continuous tracking of human faces in video sequences.
背景技术Background technique
人脸是计算机视觉系统中人机交互最便捷的方式之一。人脸检测就是在图像或图像序列中确定所有人脸的位置、大小等信息,而人脸跟踪则是持续跟踪视频序列中的一个或多个检测人脸。人脸检测与跟踪技术不仅是人脸识别、表情识别、人脸合成等技术的必要前提,而且其在智能人机交互、视频会议、智能监控、视频检索等领域有着广泛的应用价值。Human face is one of the most convenient ways of human-computer interaction in computer vision system. Face detection is to determine the position, size and other information of all faces in an image or image sequence, while face tracking is to continuously track one or more detected faces in a video sequence. Face detection and tracking technology is not only a necessary prerequisite for face recognition, expression recognition, face synthesis and other technologies, but also has a wide range of application values in intelligent human-computer interaction, video conferencing, intelligent monitoring, video retrieval and other fields.
人脸检测的重要性以及人脸模式的复杂性使得人脸检测课题一直是计算机视觉领域的一个研究热点,相关文献和方法非常多,主要分为基于启发式规则和基于统计模型两类。基于启发式规则的方法首先提取人脸的边缘、肤色、运动、对称性、轮廓、脸部器官等具有明确物理含义的特征,然后按照对人脸的先验知识制定一系列规则,最后通过检验特征是否符合这些先验规则来检测人脸。这类方法速度一般比较快,但要依赖于固定的先验规则,适应变化的能力差,虚警较多。基于统计的方法则是采用大量的“人脸”与“非人脸”样本,通过所采用的像素灰度特征或者其它变换域特征,训练和构造分类器,并利用所构造的分类器对所有可能大小的候选人脸区域进行判断,从而检测得到所有可能位置和大小的人脸。The importance of face detection and the complexity of face patterns have made face detection a research hotspot in the field of computer vision. There are many related literatures and methods, which are mainly divided into two categories based on heuristic rules and based on statistical models. The method based on heuristic rules first extracts the features with clear physical meaning such as the edge of the face, skin color, movement, symmetry, contour, facial organs, etc., then formulates a series of rules according to the prior knowledge of the face, and finally passes the test. Whether the features conform to these prior rules is used to detect faces. This type of method is generally faster, but it depends on fixed prior rules, poor ability to adapt to changes, and more false alarms. The statistical method uses a large number of "face" and "non-face" samples, trains and constructs a classifier through the pixel grayscale features or other transformation domain features, and uses the constructed classifier to classify all Candidate face regions of possible sizes are judged to detect faces of all possible positions and sizes.
尽管上述的方法是有大量样本训练得到,在统计意义上更为可靠,扩充了检测的范围,提高了检测系统的鲁棒性,较适合于复杂场景中的人脸检测,但是需要的时间比较长,实时性比较差。Although the above method is trained with a large number of samples, it is more reliable in the statistical sense, expands the detection range, improves the robustness of the detection system, and is more suitable for face detection in complex scenes, but the time required is relatively large. Long, real-time performance is relatively poor.
尽管现有技术中也采用了AdaBoost的人脸检测算法,但是基于上述方法可见,仍然存在训练时间过长、提取的人脸特征数过多的问题,同时现有人脸检测及跟踪技术在获取当前帧目标的位置时仅考虑了前一帧的直方图,将前一帧的直方图作为模板,这样容易导致一些不稳定结果,如果某帧跟踪结果不够准确,则后续帧的跟踪结果也会连续出错,还存在检测左右光照不均匀的人脸的能力差,且容易受噪声干扰,稳定性比较差的问题。Although the face detection algorithm of AdaBoost is also used in the prior art, it can be seen based on the above method that the training time is too long and the number of extracted face features is too large. The position of the frame target only considers the histogram of the previous frame, and uses the histogram of the previous frame as a template, which will easily lead to some unstable results. If the tracking result of a certain frame is not accurate enough, the tracking results of subsequent frames will also continue. If there is a mistake, there is also the problem of poor ability to detect faces with uneven illumination on the left and right, and is easily disturbed by noise, and the stability is relatively poor.
并且由于图像是视频摄像头输入的视频序列,这意味着输入摄像头的人脸存在着很大的不确定性,经常受表情、外貌的干扰,胡须、眼镜、头发也会影响人脸的外观。另外,人脸尺寸、旋转、姿态、俯仰的变化、局部区域的遮挡,以及成像条件造成的变化等,都极大地影响人脸的外观,从而影响算法性能。And because the image is a video sequence input by the video camera, this means that the face input to the camera has great uncertainty, and is often disturbed by expressions and appearance. Beards, glasses, and hair can also affect the appearance of the face. In addition, changes in face size, rotation, attitude, pitch, occlusion in local areas, and changes caused by imaging conditions all greatly affect the appearance of the face, thereby affecting the performance of the algorithm.
发明内容Contents of the invention
本发明解决现有技术中采集到的视频序列中进行人脸的实时检测与持续跟踪的方法及系统计算时间比较长、实时性比较差的技术问题,本发明的目的是采用基于AdaBoost统计分层分类器的人脸检测方法,实现正面直立人脸的实时检测,并结合基于Mean shift和直方图特征的人脸跟踪方法,实现了人脸的实时检测和跟踪,本发明的方法具有检测及跟踪迅速、实时性强的优点。The present invention solves the technical problems of the method for real-time detection and continuous tracking of human faces in the video sequences collected in the prior art and the system's relatively long calculation time and poor real-time performance. The face detection method of the classifier realizes the real-time detection of the upright face of the front, and in combination with the face tracking method based on Mean shift and histogram features, realizes the real-time detection and tracking of the face. The method of the present invention has the advantages of detection and tracking The advantages of rapidity and strong real-time performance.
本发明的目的是这样实现的:The purpose of the present invention is achieved like this:
一种视频序列中人脸的实时检测与持续跟踪的方法,包括以下步骤:A method for real-time detection and continuous tracking of human faces in a video sequence, comprising the following steps:
输入视频图像;input video image;
采用实时人脸检测算法对输入的视频图像进行人脸检测;Use the real-time face detection algorithm to detect the face of the input video image;
再采用粗细两级人脸检测算法对检测到的人脸进行检测验证;Then use the thick and thin two-level face detection algorithm to detect and verify the detected faces;
采用物体跟踪算法跟踪验证后的人脸;Use the object tracking algorithm to track the verified face;
通过对跟踪区域的验证对跟踪的人脸进行验证处理。The tracked face is verified by verifying the tracked area.
所述的实时人脸检测算法是基于AdaBoost算法由多级分类器实现。The real-time face detection algorithm is realized by a multi-level classifier based on the AdaBoost algorithm.
所述的人脸检测包含以下步骤:Described face detection comprises the following steps:
接收到的图像信息,进行图像缩放,搜索到人脸窗口;The received image information is zoomed in and out, and the face window is searched;
对检测到的人脸进行特征点定位,对人脸进行几何归一化;Perform feature point positioning on the detected face, and perform geometric normalization on the face;
对人脸进行灰度均衡化处理;Perform gray level equalization processing on the face;
对人脸进行旋转缩放;Rotate and scale the face;
得到检测的标准人脸图像。Get the detected standard face image.
所述再采用粗细两级人脸检测算法对检测到的人脸进行检测验证的步骤中包括以下情况:The step of detecting and verifying the detected faces by using the thick and thin two-level face detection algorithm includes the following situations:
如在某帧图像中检测到一个或多个人脸,在接下来的两帧图像中跟踪这些人脸,并对后续两帧图像中跟踪的人脸进行检测和验证;If one or more faces are detected in a frame of images, track these faces in the next two frames of images, and detect and verify the faces tracked in the next two frames of images;
在某位置连续三帧检测到人脸后,确定人脸存在,选择其中一个开始跟踪。After a face is detected in three consecutive frames at a certain position, it is determined that the face exists, and one of them is selected to start tracking.
本发明的方法在对人脸进行灰度均衡化处理后,还包括对人脸两侧灰度分别归一化的步骤,使人脸左右半边灰度的均值和方差相等。The method of the present invention further includes the step of normalizing the gray levels on both sides of the face after performing gray level equalization processing on the face, so that the mean value and variance of the gray levels on the left and right sides of the face are equal.
所述的人脸旋转缩放后还包括微特征计算以及分类器判决的步骤,该步骤是采用处理后的人脸图像的积分图和平方积分图得到图像中任意尺度、位置的微结构特征。After the face rotation and scaling, it also includes the steps of micro-feature calculation and classifier judgment. This step is to use the integral map and square integral map of the processed face image to obtain microstructural features of any scale and position in the image.
所述的对检测到的人脸进行验证的步骤包括:The described steps of verifying the detected faces include:
在后续帧中持续跟踪选择的人脸,如相邻帧中后一帧与前一帧的跟踪结果相似度低于设定值,停止跟踪;前一目标停止跟踪后,在后续图像中重新进行人脸检测,直到找到新的人脸,验证后进行跟踪步骤。Continue to track the selected face in the subsequent frames. If the similarity between the tracking results of the next frame and the previous frame in the adjacent frames is lower than the set value, stop tracking; after the previous target stops tracking, start again in the subsequent images. Face detection until a new face is found, followed by a tracking step after verification.
所述再采用粗细两级人脸检测算法对检测到的人脸进行检测验证的步骤中,人脸的检测分两级实现,粗检测搜索的人脸窗口的分辨率小于所述细检测搜索的人脸窗口的分辨率,对每一种尺度的图像,先将其相应缩小后,用粗检测搜索的人脸窗口进行人脸检测,淘汰非人脸窗口,再在原尺度图像中,用细检测搜索的人脸窗口对剩下的人脸候选窗口进行人脸检测。In the step of using the thick and thin two-stage face detection algorithm to detect and verify the detected faces, the face detection is realized in two stages, and the resolution of the face window of the coarse detection search is smaller than that of the fine detection search. The resolution of the face window, for each scale of the image, first reduce it accordingly, use the face window searched by the coarse detection to detect the face, eliminate the non-face window, and then in the original scale image, use the fine detection The searched face window performs face detection on the remaining face candidate windows.
在进行跟踪验证后的人脸步骤中,所述的物体跟踪算法是基于Meanshift和直方图特征做出。In the step of face tracking after verification, the object tracking algorithm is based on Meanshift and histogram features.
所述的直方图包括长期直方图、短期直方图以及颜色直方图。The histograms include long-term histograms, short-term histograms and color histograms.
对跟踪区域验证的步骤是指每隔数帧在跟踪区域进行人脸检测,如果在跟踪区域检测到正面人脸,则根据检测人脸的大小和位置更新跟踪参数,所述跟踪参数包括人脸的中心、半径和直方图特征;如果连续数百帧都跟踪到目标,但在跟踪区域都没有检测到正面人脸,停止跟踪,在后续图像中重新进行人脸检测,直到找到新的人脸,验证后进行跟踪步骤。The step of verifying the tracking area refers to performing face detection in the tracking area every few frames. If a frontal face is detected in the tracking area, the tracking parameters are updated according to the size and position of the detected face. The tracking parameters include The center, radius and histogram features of the target; if the target is tracked for hundreds of consecutive frames, but no frontal face is detected in the tracking area, stop tracking, and perform face detection in subsequent images until a new face is found , followed by a trace step after validation.
本发明还提出一种视频序列中人脸的实时检测与持续跟踪系统,包括人脸检测装置与人脸跟踪装置,人脸检测装置包括人脸处理单元、微特征计算单元及分类器单元;所述的人脸处理单元接收到图像信息,对接收到的图像进行缩放,穷举搜索候选人脸窗口,计算窗口灰度的均值和方差;所述的特征计算单元根据AdaBoost算法计算出各个窗口的微结构特征,并将其传送给所述的分类器单元进行判决,分类器单元采用粗细两级人脸检测算法进行判决后将其传送给人脸跟踪装置;所述的人脸跟踪装置包括物体跟踪单元和跟踪区域验证单元,物体跟踪单元采用直方图特征进行计算,实现对图像跟踪,跟踪区域验证单元对跟踪的图像进行区域检测,对跟踪的人脸进行验证。The present invention also proposes a real-time detection and continuous tracking system for human faces in video sequences, including a human face detection device and a human face tracking device, and the human face detection device includes a human face processing unit, a micro-feature calculation unit and a classifier unit; The face processing unit described above receives the image information, zooms the received image, exhaustively searches the candidate face window, and calculates the mean value and variance of the window gray scale; Microstructural features, and send it to the described classifier unit for judgment, the classifier unit adopts thick and thin two-level face detection algorithm to make a judgment and then send it to the human face tracking device; the described human face tracking device includes object The tracking unit and the tracking area verification unit, the object tracking unit uses the histogram feature for calculation to realize image tracking, the tracking area verification unit performs area detection on the tracked image, and verifies the tracked face.
所述的人脸检测装置的分类器单元还包含粗细两级人脸检测单元,接收微结构特征,进行粗细两级检测:粗检测搜索的人脸窗口的分辨率小于所述细检测搜索的人脸窗口的分辨率,对每一种尺度的图像,先将其相应缩小后,用粗检测搜索的人脸窗口进行人脸检测,淘汰非人脸窗口,再在原尺度图像中,用细检测搜索的人脸窗口对剩下的人脸候选窗口进行人脸检测,确定人脸窗口。The classifier unit of the described face detection device also includes a thick and thin two-level face detection unit, which receives microstructure features and performs thick and thin two-level detection: the resolution of the human face window searched by the coarse detection is smaller than that of the human face searched by the fine detection. For the resolution of the face window, for each scale image, first reduce it accordingly, use the face window searched by coarse detection for face detection, eliminate non-face windows, and then use fine detection search in the original scale image Face detection is performed on the remaining face candidate windows to determine the face window.
所述物体跟踪单元采用的直方图包括长期直方图、短期直方图以及颜色直方图。The histograms used by the object tracking unit include long-term histograms, short-term histograms and color histograms.
本发明产生的技术效果是显著的:The technical effect that the present invention produces is remarkable:
本发明所述的方法和系统是基于AdaBoost算法进行检测的,并采用粗模型和细模型进行人脸检测验证,使本发明取得非常高的检测率,且检测速度非常快,效率高,可实时实现;本发明在检测过程中根据标准人脸灰度的均值和方差对左右人脸分辨率进行灰度的归一化,消除左右人脸光照不均匀的影响;本发明在人脸检测完毕后在短时间内继续进行人脸的跟踪和验证,消除人脸检测虚警的影响;人脸跟踪过程中引入长期直方图和短期直方图两个局部特征,反映前面图像中目标直方图的变化过程,保证算法可以跟踪姿态不断变化的运动目标。The method and system of the present invention are based on the AdaBoost algorithm for detection, and the coarse model and the fine model are used for face detection and verification, so that the present invention can achieve a very high detection rate, and the detection speed is very fast, the efficiency is high, and real-time Realize; the present invention carries out the normalization of grayscale to left and right people's face resolution according to the mean value and the variance of standard people's face grayscale in detection process, eliminates the influence of uneven illumination of left and right people's faces; The present invention is after people's face detection finishes Continue to track and verify faces in a short period of time to eliminate the influence of false alarms in face detection; introduce two local features of long-term histogram and short-term histogram in the process of face tracking to reflect the change process of the target histogram in the previous image , to ensure that the algorithm can track the moving target whose attitude changes constantly.
附图说明Description of drawings
图1为本发明的人脸检测与跟踪方法的流程图。FIG. 1 is a flow chart of the face detection and tracking method of the present invention.
图2a和图2b为本发明的人脸检测与跟踪方法的检测及跟踪结果的示意图。2a and 2b are schematic diagrams of the detection and tracking results of the face detection and tracking method of the present invention.
图3为基于AdaBoost分层分类器实现的人脸检测流程示意图。Figure 3 is a schematic diagram of the face detection process based on the AdaBoost hierarchical classifier.
图4为分类器中的部分正样本人脸图像。Figure 4 shows some positive sample face images in the classifier.
图5为分类器中部分不包含人脸的反样本图像。Figure 5 is a part of the negative sample images that do not contain faces in the classifier.
图6a至图6d为人脸样本的标定与采集的示意图。6a to 6d are schematic diagrams of calibration and collection of face samples.
图7为经过两侧灰度归一化处理后的样本。Figure 7 is a sample after gray normalization processing on both sides.
图8为原始样本和经过镜像、左右旋转、放大处理后的样本。Figure 8 shows the original sample and the sample after mirroring, left and right rotation, and zooming in.
图9为尺度归一化后的反样本数据。Figure 9 shows the scale normalized negative sample data.
图10为本发明的人脸检测算法的实施例选择的七组微特征。Fig. 10 shows seven groups of micro-features selected by the embodiment of the face detection algorithm of the present invention.
图11为AdaBoost训练算法第K层分类器的训练流程示意图。Fig. 11 is a schematic diagram of the training process of the K-th layer classifier of the AdaBoost training algorithm.
图12a和12b为人脸检测后处理结果一个实施例的示意图。12a and 12b are schematic diagrams of an embodiment of processing results after face detection.
图13a和图13b为本发明的人脸跟踪示意图。Fig. 13a and Fig. 13b are schematic diagrams of face tracking in the present invention.
图14为人脸检测结果示意图。FIG. 14 is a schematic diagram of face detection results.
图15为人脸检测与跟踪结果的示意图。Fig. 15 is a schematic diagram of face detection and tracking results.
图16为本发明的系统一种构成框图。Fig. 16 is a block diagram of the system of the present invention.
图17为本发明的系统的另一种构成框图。Fig. 17 is another block diagram of the system of the present invention.
具体实施方式Detailed ways
本发明提出一种在视频序列中人脸的实时检测与持续跟踪的方法,结合图1所示的内容:The present invention proposes a method for real-time detection and continuous tracking of faces in video sequences, combined with the content shown in Figure 1:
首先,输入视频图像,在该步骤中由摄像头实时输入视频图像;At first, input video image, in this step, input video image by camera in real time;
然后,采用实时人脸检测算法对输入的视频图像进行人脸检测;Then, the real-time face detection algorithm is used to detect the face of the input video image;
接收到实时输入的视频图像后,对每帧图像进行搜索,检测人脸的存在;如图2a所示,给出了人脸检测的结果,其中正方形框表示为检测到的人脸;在检测过程中,如果在某帧图像中检测到一个或多个人脸,则在接下来的两帧图像中跟踪这些人脸,并对后续两帧图像中跟踪的人脸进行检测和验证,判断前面的检测结果是否是真的人脸;After receiving the real-time input video image, search each frame of image to detect the existence of human face; as shown in Figure 2a, the result of human face detection is given, where the square box represents the detected human face; in the detection In the process, if one or more faces are detected in a frame of images, these faces will be tracked in the next two frames of images, and the faces tracked in the next two frames of images will be detected and verified, and the previous Whether the detection result is a real face;
只有在某个位置连续三帧都检测到人脸后,算法才认为该位置人脸存在,如果场景中存在有多个人脸,挑选出一个人脸开始跟踪;本发明的实施例中采用跟踪最大的人脸进行说明;Only when a human face is detected in a certain position for three consecutive frames, the algorithm considers that the human face exists at this position. If there are many human faces in the scene, a human face is selected to start tracking; face description;
在后续帧中持续跟踪挑选出来的人脸,如果相邻帧中后一帧与前一帧的跟踪结果的相似度低于设定值(该设定值可以任意设定),则停止跟踪;如果某个跟踪目标所在区域长时间未检测到正面直立人脸,则认为该目标的跟踪价值不大,停止跟踪该目标;前一个跟踪目标停止跟踪后,在后续图像中重新进行人脸检测,直到找到新的人脸,跟踪新的人脸。如图2b给出了人脸跟踪的结果。Continue to track the selected faces in the subsequent frames, if the similarity between the tracking results of the next frame and the previous frame in the adjacent frames is lower than the set value (the set value can be set arbitrarily), then stop tracking; If no frontal upright face is detected in the area where a tracking target is located for a long time, it is considered that the tracking value of the target is not great, and the tracking of the target is stopped; after the previous tracking target stops tracking, face detection is performed again in subsequent images, Until a new face is found, track the new face. Figure 2b shows the results of face tracking.
以下结合样本的训练过程,对于本发明的检测过程进行说明,本发明是采用统计训练方法进行场景中正面人脸的检测,并采用AdaBoost理论实现人脸检测统计模型的训练。基于AdaBoost的人脸检测算法首先由大量“人脸”和“非人脸”样本训练一个“人脸/非人脸”二类分类器,在检测过程中,由该分类器确定某个尺度的矩形窗口是否是人脸,设矩形长为m,宽为n,则人脸检测的流程就是:首先是按照一定比例连续放缩图像,在得到的系列图像中穷举搜索和判别所有大小m×n像素窗口,将各个窗口输入到“人脸/非人脸”分类器中,留下识别为人脸的候选窗口,再采用相邻位置的候选,对合并后的平均值进行计算,输出所有检测到的人脸的位置、大小等信息。The following describes the detection process of the present invention in conjunction with the training process of the samples. The present invention uses the statistical training method to detect the front face in the scene, and uses the AdaBoost theory to realize the training of the statistical model of human face detection. The face detection algorithm based on AdaBoost first trains a "face/non-face" classifier from a large number of "face" and "non-face" samples. Whether the rectangular window is a human face or not, let the length of the rectangle be m and the width be n, then the process of face detection is: firstly, the image is continuously zoomed in according to a certain ratio, and all the obtained series of images are exhaustively searched and identified with a size of m× n-pixel window, input each window into the "face/non-face" classifier, leave the candidate window recognized as a face, and then use the candidate of the adjacent position to calculate the combined average value, and output all detections The location, size and other information of the received face.
本发明所述的一种实时人脸检测算法,是使用一种类似Harr小波的微结构特征来表达人脸模式,并结合上述的AdaBoost算法提出了一种特征选择方法,将多个基于单个特征的弱分类器组成为一个强分类器,然后将多个强分类器级联成一个完整的人脸检测分类器,结合图3所示。本发明中检测过程中形成的每层分类器都是由AdaBoost算法训练得到的一个强分类器,每层强分类器又由一定数量的弱分类器组成。检测的时候,如果某一层强分类器判别一个子窗口为False就排除此子窗口而不进行进一步的判别,如果输出为True则使用下一层更复杂的分类器对子窗口进行判别。在候选窗口的搜索过程中,每一层强分类器都能让几乎全部人脸样本通过,而拒绝大部分非人脸样本。这样输入低层强分类器的窗口就多,而输入高层的窗口大大减少。A kind of real-time human face detection algorithm described in the present invention is to use a kind of microstructural feature similar to Harr wavelet to express human face mode, and proposes a kind of feature selection method in conjunction with above-mentioned AdaBoost algorithm, multiple based on single feature The weak classifiers are composed into a strong classifier, and then multiple strong classifiers are cascaded into a complete face detection classifier, as shown in Figure 3. Each layer of classifiers formed in the detection process in the present invention is a strong classifier trained by the AdaBoost algorithm, and each layer of strong classifiers is composed of a certain number of weak classifiers. When detecting, if a strong classifier of a certain layer judges a sub-window as False, the sub-window will be excluded without further discrimination. If the output is True, the next layer of more complex classifiers will be used to discriminate the sub-window. During the search process of candidate windows, each layer of strong classifiers can pass almost all face samples and reject most non-face samples. In this way, there are more windows input to the low-level strong classifier, while the windows input to the high-level are greatly reduced.
人脸检测算法需要对穷举搜索后的图像放缩一定尺度,得到人脸窗口,在本实施例中,以320×240像素的图像为例,如按照1.25的比例缩小10次,并以20×20大小的窗口逐个像素进行搜索,共需判断170000余个窗口。这意味着一次人脸检测都需要搜索大量的窗口,面对如此大的计算量,图3的检测流程中低层分类器必须非常简单,即前面的分类器包含的弱分类器数量少。后面的分类器复杂,包含的弱分类器数量多,这样后面的层可采用较多的特征来拒除与人脸相似的候选窗口,从而保证了较低的误检率。The face detection algorithm needs to scale the image after the exhaustive search to a certain scale to obtain the face window. In this embodiment, take the image of 320×240 pixels as an example, such as reducing it 10 times according to the ratio of 1.25, and using 20 A window of size ×20 is searched pixel by pixel, and a total of more than 170,000 windows need to be judged. This means that a face detection needs to search a large number of windows. Faced with such a large amount of calculation, the low-level classifiers in the detection process in Figure 3 must be very simple, that is, the number of weak classifiers contained in the previous classifiers is small. The latter classifier is complex and contains a large number of weak classifiers, so that the latter layer can use more features to reject candidate windows similar to human faces, thereby ensuring a lower false detection rate.
在进行人脸检测时,将检测到的图像与训练的样本数据进行对比,其中进行模型的训练时,需要收集大量的人脸正样本和反样本数据,通过这些数据的收集,建立本发明的样本数据库,该数据库中的正样本包含多幅人脸样本,这些样本包括不同表情、不同肤色、不同年龄的人脸,包含-20°到20°深度旋转的人脸,包含戴与不戴眼镜人脸,部分样本如图4所示。反样本数据就是大量不包含人脸的图像,包括风景图像、动物、文字等,参见图5所示的内容。在训练的时候分析所有人脸的关键特征点,确定每个正样本人脸的两眼中心、鼻尖、嘴巴中心以及下巴。根据这些标定点对各个人脸进行几何归一化,即将人脸图像的主要器官位置矫正到标准位置,减小样本间的尺度、平移和平面旋转差异,然后根据器官位置剪裁出人脸区域成为人脸样本,使人脸样本尽量少地引入背景干扰,且不同人脸样本的器官位置具有一致性。在进行人脸检测的过程中,也需要对检测到的人脸作出器官位置的确定,以进行人脸的几何归一化。When performing face detection, the detected image is compared with the training sample data. When performing model training, a large number of face positive samples and negative sample data need to be collected. Through the collection of these data, the method of the present invention is established. Sample database, the positive samples in this database contain multiple face samples, these samples include faces with different expressions, different skin colors, and different ages, including faces with depth rotations from -20° to 20°, including those with and without glasses Some samples of human faces are shown in Figure 4. The negative sample data is a large number of images that do not contain faces, including landscape images, animals, text, etc., see the content shown in Figure 5. During training, analyze the key feature points of all faces, and determine the center of the two eyes, the tip of the nose, the center of the mouth, and the chin of each positive sample face. Geometrically normalize each face according to these calibration points, that is, correct the main organ position of the face image to the standard position, reduce the scale, translation and plane rotation differences between samples, and then cut out the face area according to the organ position as Face samples, so that the face samples introduce as little background interference as possible, and the organ positions of different face samples are consistent. In the process of face detection, it is also necessary to determine the position of the organs of the detected face for geometric normalization of the face.
参考图6a~图6d所示的内容,表示了一幅标准人脸图像对各个人脸样本进行几何归一化和人脸区域的裁剪过程:首先确定前面提供的检测窗口尺度m×n为20×20,接着获取一幅标准的正面人脸图像,标准图像中两眼的y坐标一致,人脸也完全对称,如图6a,标定该图像的五个关键特征点。根据该图像中双眼的距离和位置确定裁剪的正方形人脸区域的位置。设两眼的距离为r,两眼连线的中心点为(xcenter,ycenter),采集矩形的长宽设为2r,即两倍双眼间距,则矩形裁剪区域的坐标(xleft,ytop,xright,ybottom)为:Referring to the contents shown in Figures 6a to 6d, it shows a standard face image to perform geometric normalization on each face sample and the process of clipping the face area: first, determine that the detection window scale m×n provided earlier is 20 ×20, and then obtain a standard frontal face image. In the standard image, the y coordinates of the two eyes are consistent, and the face is completely symmetrical, as shown in Figure 6a, and the five key feature points of the image are calibrated. The position of the cropped square face area is determined according to the distance and position of the two eyes in the image. Suppose the distance between the two eyes is r, the center point of the connecting line between the two eyes is (x center , y center ), and the length and width of the acquisition rectangle are set to 2r, which is twice the distance between the eyes, then the coordinates of the rectangular clipping area (x left , y center top , x right , y bottom ) are:
将裁剪的人脸区域归一化到20×20的尺寸,如图6b,并获取归一化后五个标定点的坐标[xstad(i),ystad(i)],i=0,1,2,3,4。Normalize the cropped face area to a size of 20×20, as shown in Figure 6b, and obtain the coordinates [x stad (i), y stad (i)] of five calibration points after normalization, i=0, 1, 2, 3, 4.
任意给定原始人脸样本和标定的五个特征点[xlabel(i),ylabel(i)],i=0,1,2,3,4,如图6c,计算这五个点与标准图像归一化后五点坐标间的仿射变换系数。在变换过程中应该保证各个人脸样本总的形状不变,即长脸仍应该是长脸,短脸仍应该是短脸,这样检测算法就可以检测不同类型的人脸,因此仿射变换式中不能加入人脸各个方向的拉伸变换,我们仅考虑旋转和整体缩放两个变换,由此可以确定仿射变换公式为:Any given original face sample and five marked feature points [x label (i), y label (i)], i=0, 1, 2, 3, 4, as shown in Figure 6c, calculate these five points and The affine transformation coefficient between the five coordinates of the standard image after normalization. During the transformation process, it should be ensured that the overall shape of each face sample remains unchanged, that is, long faces should still be long faces, and short faces should still be short faces, so that the detection algorithm can detect different types of faces, so the affine transformation formula cannot Adding the stretch transformation in all directions of the face, we only consider the two transformations of rotation and overall scaling, so we can determine the affine transformation formula as:
采用最小二乘法可以获取上式中的变换参数(a,b,c,d)。The transformation parameters (a, b, c, d) in the above formula can be obtained by using the least square method.
设裁剪后的人脸图像为I,该图像尺寸为20×20,则由仿射变换系数可以计算该图像中任意一点(x,y)到原始样本中对应点坐标(xori,yori)。Let the cropped face image be I, and the size of the image is 20×20, then the coordinates (x ori , y ori ) of any point (x, y) in the image to the corresponding point in the original sample can be calculated by the affine transformation coefficient .
为消除噪声的影响,裁剪后的图像中(x,y)的像素值设为对应点(xori,yori)邻域范围内所有点像素值的均值。由此可以获取I中所有点的像素值,如图6d。In order to eliminate the influence of noise, the pixel value of (x, y) in the cropped image is set to the mean value of the pixel values of all points within the neighborhood of the corresponding point (x ori , y ori ). Thus, the pixel values of all points in I can be obtained, as shown in Figure 6d.
在进行检测的过程中,由于外界光照、成像设备等因素可能导致人脸图像亮度或对比度异常,出现强阴影或反光等情况,另外不同人种的肤色间也存在这差异,因此需要对几何归一化后的人脸样本进行灰度均衡化处理,改善其灰度分布,增强模式间的一致性。由于在人脸检测过程中,每一个搜索窗口都需要进行灰度均衡化处理,因此不能采用计算较复杂的方法进行灰度归一化。本发明采用可用快速算法实现的图像灰度均值、方差归一化进行样本的灰度均衡化。另外,受不同方向特别是侧面光照的影响,实际场景中左右两边人脸的亮度经常存在着明显的差异。因此本发明对两侧人脸灰度分别进行归一化,使人脸左右半边灰度的均值和方差都等于一个设定的标准值。图7给出了部分经过两侧归一化处理后的人脸图像。In the process of detection, due to factors such as external lighting and imaging equipment, the brightness or contrast of the face image may be abnormal, and strong shadows or reflections may appear. In addition, there are differences in the skin color of different races. The normalized face samples are subjected to gray level equalization processing to improve their gray level distribution and enhance the consistency between modes. Since in the process of face detection, each search window needs to perform gray level equalization processing, it is not possible to use a computationally complex method for gray level normalization. The invention adopts the image gray mean value and variance normalization which can be realized by a fast algorithm to carry out the gray level equalization of the sample. In addition, due to the influence of different directions, especially the side lighting, there are often obvious differences in the brightness of the left and right faces in the actual scene. Therefore, the present invention normalizes the gray scales of the faces on both sides, so that the mean value and variance of the gray scales of the left and right halves of the face are both equal to a set standard value. Figure 7 shows some face images after normalization processing on both sides.
为增强分类器对人脸一定角度的旋转和尺寸变化的检测鲁棒性,本发明对每个样本进行镜像变换、旋转±20°角度、尺寸放大1.1倍,如图8。每个样本扩充为五个样本,如此组成了AdaBoost训练的正样本集。In order to enhance the robustness of the classifier to detect face rotation and size changes at a certain angle, the present invention performs mirror transformation on each sample, rotates at an angle of ±20°, and enlarges the size by 1.1 times, as shown in Figure 8. Each sample is expanded to five samples, thus forming the positive sample set for AdaBoost training.
反样本数据则在AdaBoost的每层训练过程中到反样本图像库中随机选取。先随机确定反样本图像的序号,随机确定反样本的尺寸和位置,接着到该序号对应图像中裁剪出对应的区域,将裁剪图像归一化到20×20的尺寸,得到一个反样本,图9是部分反样本数据。The counter-sample data is randomly selected from the counter-sample image library during the training process of each layer of AdaBoost. First randomly determine the serial number of the counter-sample image, randomly determine the size and position of the counter-sample, then cut out the corresponding area in the image corresponding to the serial number, and normalize the cropped image to a size of 20×20 to obtain a counter-sample, as shown in Fig. 9 is part of the counter-sample data.
本发明在进行特征计算的时候采用了特征提取方法,在本实施例中使用了七组微特征,来有效地表达人脸模式的结构特点。图10给出了20×20图像中所有的微特征结构,计算图像中对应黑色区域和白色区域内像素灰度均值的差值得到特征。前面六组微特征中黑色矩形和白色矩形的大小保持一致,而第七组微特征中白色矩形的长宽是黑色矩形的三倍。各组微特征中黑色矩形或者白色矩形的长、宽可以任意选择,即每个矩形的尺寸可选1到20的任意数值。各组微特征中中心点的位置也可任意选择,因此理论上在20×20图像中可以获取20×20×20×20×7=1120000个特征。考虑到很多组特征中黑色或者白色区域已经到了20×20图像外面,对这类特征我们忽略不计。因此有效的特征数为89199。The present invention adopts a feature extraction method when performing feature calculation. In this embodiment, seven groups of micro-features are used to effectively express the structural characteristics of the human face pattern. Figure 10 shows all the micro-feature structures in the 20×20 image, and the feature is obtained by calculating the difference between the pixel gray value corresponding to the black area and the white area in the image. The size of the black rectangle and the white rectangle in the first six groups of microfeatures remain the same, while the length and width of the white rectangle in the seventh group of microfeatures are three times that of the black rectangle. The length and width of the black rectangle or the white rectangle in each group of micro-features can be selected arbitrarily, that is, the size of each rectangle can be any value from 1 to 20. The position of the central point in each group of micro-features can also be selected arbitrarily, so theoretically 20×20×20×20×7=1120000 features can be obtained in a 20×20 image. Considering that the black or white area in many groups of features has reached outside the 20×20 image, we ignore such features. So the effective number of features is 89199.
在人脸检测过程中需要不断的计算微特征,并将微特征输入到各层AdaBoost强分类器中进行判决。因此微特征计算效率就决定了人脸检测算法的效率。可以利用整幅图像的积分图和平方积分图快速得到图像中任意尺度、任意位置的一种微结构特征,从而为人脸实时检测系统的实现提供了可能,而且采用这种方法无需对抽取的各个20×20窗口的所有像素值进行灰度归一化,只需结合20×20窗口左右半边的灰度均值和方差进行微结构特征的归一化。In the process of face detection, it is necessary to continuously calculate the micro-features, and input the micro-features into each layer of AdaBoost strong classifier for judgment. Therefore, the micro-feature calculation efficiency determines the efficiency of the face detection algorithm. A microstructural feature of any scale and any position in the image can be quickly obtained by using the integral map and square integral map of the entire image, which provides the possibility for the realization of a real-time face detection system, and this method does not require All pixel values in the 20×20 window are normalized to grayscale, and only the grayscale mean and variance of the left and right half of the 20×20 window are combined to normalize the microstructural features.
如图11所示,是AdaBoost算法训练流程中第K层分类器的训练过程。首先将所有归一化后的正样本人脸数据输入到前K-1层已经训练好的分类器中,将通过这些分类器的正样本输入到第K层分类器的训练模块中。再到前面提到的5400幅反样本图像中随机选择反样本,将各个反样本也输入到前K-1层分类器中,将通过这K-1层分类器的样本作为第K层分类器的训练模块的输入反样本。在训练过程中我们会保证每层分类器不会淘汰或淘汰非常少的正样本人脸。因此最后输入到训练模块中的正样本基本保持在一定的数量。为获取较好的分类性能,且训练效率不至过低,本发明选择与正样本数量相当的反样本。因此当随机选择的反样本数达到正样本相当的数量时停止反样本的选取,开始第K层分类器的训练。As shown in Figure 11, it is the training process of the K-th layer classifier in the AdaBoost algorithm training process. First, all the normalized positive sample face data are input into the classifiers that have been trained in the first K-1 layer, and the positive samples that pass these classifiers are input into the training module of the K-th layer classifier. Then randomly select the anti-sample from the 5,400 anti-sample images mentioned above, and input each anti-sample to the first K-1 layer classifier, and use the samples that passed the K-1 layer classifier as the K-th layer classifier The input counter-sample of the training module of . During the training process, we will ensure that each classifier will not eliminate or eliminate very few positive sample faces. Therefore, the final positive samples input into the training module are basically kept at a certain number. In order to obtain better classification performance and keep the training efficiency from being too low, the present invention selects negative samples equal to the number of positive samples. Therefore, when the number of randomly selected negative samples reaches the equivalent number of positive samples, the selection of negative samples is stopped, and the training of the K-th layer classifier is started.
AdaBoost的强分类器由基于单个特征的弱分类器组成,即每个弱分类器对应一个微特征。本发明的弱分类器定义为:AdaBoost's strong classifier consists of weak classifiers based on a single feature, that is, each weak classifier corresponds to a micro-feature. The weak classifier of the present invention is defined as:
其中x是20×20的图像窗口,gj(x)表示图像在第j个特征下的特征值,θj是第j个特征对应的判决阈值,hj(x)表示图像在第j个特征下的判决输出。上式为每个弱分类器定义了一个微特征j和阈值θj,判决方式有三种可能,即根据输入的微特征gj(x)是大于该阈值,还是小于该阈值,还是绝对值小于该阈值决定判决输出是1还是0。每个弱分类器只能选一种判决可能,在训练过程中可根据所有的正反样本对当前微特征进行处理,分别获取三种判决对应的阈值及训练样本集上的分类错误率,将错误率最小的方式作为当前弱分类器的判决方式。Where x is a 20×20 image window, g j (x) represents the feature value of the image under the j feature, θ j is the decision threshold corresponding to the j feature, h j (x) represents the image under the j feature The decision output under the feature. The above formula defines a micro-feature j and a threshold θ j for each weak classifier. There are three possible judgment methods, that is, according to whether the input micro-feature g j (x) is greater than the threshold, or less than the threshold, or the absolute value is less than This threshold determines whether the decision output is 1 or 0. Each weak classifier can only choose one decision possibility. During the training process, the current micro-features can be processed according to all the positive and negative samples, and the thresholds corresponding to the three kinds of decisions and the classification error rate on the training sample set can be obtained respectively. The method with the smallest error rate is used as the judgment method of the current weak classifier.
接着对每层强分类器进行训练,而训练过程实际上就是获取各个弱分类器对应的微特征序号的过程,从而使各弱分类器在训练集上总的分类能力最强。我们设定每层分类器包含的弱分类器固定,设为T,且一个微特征只能出现一次。第一个分类器对应的微特征在后面的训练过程中将不考虑。下面列出了算法的训练流程:Then, each layer of strong classifiers is trained, and the training process is actually the process of obtaining the micro-feature serial numbers corresponding to each weak classifier, so that the overall classification ability of each weak classifier on the training set is the strongest. We set the weak classifiers contained in each classifier to be fixed, set to T, and a micro-feature can only appear once. The micro-features corresponding to the first classifier will not be considered in the subsequent training process. The training process of the algorithm is listed below:
给定包含n个样本的训练集{(xi,yi)},i=0,1,...,n-1{()},yi=0或1,表示对应的输入样本xi是人脸样本还是非人脸样本,其中人脸样本数量为m,非人脸样本数量为1; Given a training set {(x i , y i )} containing n samples, i=0, 1, ..., n-1{()}, y i =0 or 1, representing the corresponding input sample x i is a face sample or a non-face sample, where the number of face samples is m, and the number of non-face samples is 1;
选择误分类风险倍数c,表示训练样本分类错误的风险大小,则对于人脸样本
迭代次数t=0,1,....T-1:(T即为希望选择的弱分类器的个数) The number of iterations t=0, 1,...T-1: (T is the number of weak classifiers to be selected)
(1)对每个特征j,利用单个特征训练分类器hj,根据训练样本集的权值Wt得到最优的阈值参数,使得hj的错误率εj最小:
(2)得到错误率最小的弱分类器作为当前层强分类器的第t个弱分类器ht,对应的特征序号为ft,对应的错误率为εt;(2) Get the weak classifier with the smallest error rate as the t-th weak classifier h t of the current layer strong classifier, the corresponding feature number is f t , and the corresponding error rate is ε t ;
(3)计算参数
(4)更新所有样本的权重
输出最后的强分类器: 强分类器输出为1 Output the final strong classifier: Strong classifier output is 1
表示输入样本x通过该层强分类器,否则认为输入样本是非人脸。Indicates that the input sample x passes the strong classifier of this layer, otherwise the input sample is considered to be a non-face.
每层强分类器对正样本的分类错误率要尽可能的低,对单层强分类器而言,c越大,正样本的初始权重越大,反样本的初始权重越小,训练算法为降低各个弱分类器的错误率εt,会尽可能保证权重大的样本分类正确,因此c越大,正样本的分类错误率就越小,反样本的分类错误率就越高,因此我们在训练过程采用手动调节参数c的方法,使各层强分类器对训练正样本的分类错误率非常小,一般要小于0.05%。而对单层分类器而言,我们对反样本的分类正确率即虚警率则没有明确的要求。因为人脸检测算法由多个分类器级联而成,每层的虚警率并不需要太低,而数十层的虚警率相乘得到的总虚警率就可以非常小,如果每层的正样本错误率都小于0.05%,总的正确率仍可达到99%,这样就可保证算法可以检测各种类型的训练样本,又训练样本已经包括了各种类型、多个方向、多个标准和姿态的人脸,因此最后实现的人脸检测模型可以检测加入多类干扰的正面人脸样本。The classification error rate of each layer of strong classifiers for positive samples should be as low as possible. For a single-layer strong classifier, the larger c is, the larger the initial weight of positive samples is, and the smaller the initial weight of negative samples is. The training algorithm is Reducing the error rate ε t of each weak classifier will ensure that the samples with heavy weights are classified correctly as much as possible. Therefore, the larger c is, the smaller the classification error rate of positive samples is, and the higher the classification error rate of negative samples is. Therefore, we have The training process adopts the method of manually adjusting the parameter c, so that the classification error rate of each layer of strong classifiers for training positive samples is very small, generally less than 0.05%. For single-layer classifiers, we have no clear requirements for the classification accuracy rate of negative samples, that is, the false alarm rate. Because the face detection algorithm is formed by cascading multiple classifiers, the false alarm rate of each layer does not need to be too low, and the total false alarm rate obtained by multiplying the false alarm rates of dozens of layers can be very small. The positive sample error rate of the layer is less than 0.05%, and the total correct rate can still reach 99%, which can ensure that the algorithm can detect various types of training samples, and the training samples already include various types, multiple directions, multiple Therefore, the final implemented face detection model can detect positive face samples with multiple types of interference.
另外,单层强分类器的弱分类器数T也需要仔细调整。T越大,弱分类器数量越多,虚警率一般越小;由于将各个候选人脸窗口输入到强分类器中时,需要先计算该层强分类器对应的T个微特征,因此T越大也意味着该层强分类器的计算效率偏高。我们在训练过程中需要根据虚警率不断调节各层分类器T的大小,在虚警率与计算效率间寻找折衷,原则是每层分类器的虚警率要小,同时T不能太大,计算效率要比较高。In addition, the number T of weak classifiers of single-layer strong classifiers also needs to be carefully adjusted. The larger T is, the larger the number of weak classifiers is, and the lower the false alarm rate is generally; since each candidate face window is input into the strong classifier, T micro-features corresponding to the strong classifier of this layer need to be calculated first, so T A larger value also means that the computational efficiency of the strong classifier at this layer is higher. During the training process, we need to continuously adjust the size of the classifier T of each layer according to the false alarm rate, and find a compromise between the false alarm rate and computational efficiency. The principle is that the false alarm rate of each classifier should be small, and T should not be too large. Computational efficiency is higher.
在本发明的人脸检测中,采用粗细两级人脸检测算法对检测到的人脸进行检测验证;以下结合具体实施例进行说明:In the face detection of the present invention, a two-level face detection algorithm is used to detect and verify the detected faces; the following will be described in conjunction with specific embodiments:
由于20×20图像的微特征数达到了89199,这意味着在每次弱分类器的训练过程中都需要从89199特征中搜索出最佳的特征,这个过程是非常耗时的。为了提高训练算法的效率,同时也为了提高检测算法的性能,本发明提出的粗细两级人脸检测算法,将人脸检测过程分两级实现,粗检测搜索的人脸窗口尺寸固定为10×10,细检测搜索的人脸窗口尺寸才是20×20。在检测过程中我们同样获取不同尺度的图像,接着对各个尺度图像的分辨率再减半,搜索大小为10×10的人脸候选窗口,将各个窗口输入到粗检测的各层强分类器中,计算各层强分类器的微特征,并进行判决,淘汰非人脸窗口,最后剩下少数候选人脸窗口输入到细检测中;将通过粗检测的10×10窗口扩为20×20窗口,进行细检测,到原来的分辨率未减半的尺度图像中继续搜索这些候选窗口,确定最终的人脸窗口。同样,训练过程也分为两部分,先训练粗检测模型,再训练细检测模型。Since the number of micro-features of a 20×20 image reaches 89199, it means that the best features need to be searched out from the 89199 features in each weak classifier training process, and this process is very time-consuming. In order to improve the efficiency of the training algorithm, and also in order to improve the performance of the detection algorithm, the thick and thin two-stage face detection algorithm proposed by the present invention implements the face detection process in two stages, and the face window size of the coarse detection search is fixed at 10× 10. The size of the face window for fine detection and search is 20×20. In the detection process, we also obtain images of different scales, and then halve the resolution of images of each scale, search for face candidate windows with a size of 10×10, and input each window into the strong classifiers of each layer for rough detection , calculate the micro-features of the strong classifiers at each layer, and make a judgment, eliminate the non-face windows, and finally a few candidate face windows are input into the fine detection; expand the 10×10 window that passed the rough detection to a 20×20 window , perform fine detection, and continue to search for these candidate windows in the scale image whose resolution has not been halved, and determine the final face window. Similarly, the training process is also divided into two parts, first training the coarse detection model, and then training the fine detection model.
将训练集中所有的人脸正样本分辨率再减半,得到10×10的粗检测人脸样本。而10×10人脸样本总的微特征数仅为5676,这样粗检测模型的训练效率就远高于细检测模型的效率。另外10×10图像中人脸各个微特征的计算效率也高于20×20图像中微特征的效率,因此采用这种两级检测方法也可以大大提高人脸检测的速度。The resolution of all positive face samples in the training set is halved again to obtain 10×10 rough detection face samples. However, the total number of micro-features in a 10×10 face sample is only 5676, so the training efficiency of the coarse detection model is much higher than that of the fine detection model. In addition, the calculation efficiency of each micro-feature of the face in the 10×10 image is also higher than that of the micro-features in the 20×20 image, so this two-stage detection method can also greatly improve the speed of face detection.
任意输入一幅图像,为检测该图像中一定尺度范围的人脸,我们分尺度对该图像进行缩放。例如,如果要检测320×240图像中尺度从20×20到240×240的人脸,我们需要在多个尺度缩小图像。对粗检测而言,最小缩小倍数应为2,最大缩小倍数为24;对细检测而言,最小缩小倍数为1,最大缩小倍数为12。如果将相邻两尺度的缩小倍数的比例设为1.25,则图像的缩小倍数分别为(对粗检测而言)2、2.5、3.13、3.91、4.88、6.10、7.63、9.54、11.9、14.9、18.6、23.3,共12个尺度。Arbitrarily input an image, in order to detect faces in a certain scale range in the image, we scale the image by scale. For example, if we want to detect faces with scales from 20×20 to 240×240 in a 320×240 image, we need to downscale the image at multiple scales. For coarse detection, the minimum reduction factor should be 2, and the maximum reduction factor should be 24; for fine detection, the minimum reduction factor should be 1, and the maximum reduction factor should be 12. If the ratio of the reduction multiples of two adjacent scales is set to 1.25, the reduction multiples of the image are (for rough detection) 2, 2.5, 3.13, 3.91, 4.88, 6.10, 7.63, 9.54, 11.9, 14.9, 18.6 , 23.3, a total of 12 scales.
将所有尺度图像中通过粗细两级检测的正方形框反变换到原始输入图像的尺度和位置,得到原始图像中人脸的候选位置和候选尺寸。一般情况下,一张真实人脸往往会在不同的尺度下和相邻的位置处检测出多次,而虚警的出现往往比较孤立,图12a是一个示例。此时需要对检测结果进行后处理,将相邻位置的人脸框合并,如果两个候选人脸框的大小差异和位置差异都非常小,或者这两个人脸框重叠面积非常大,就可以将这两个框合二为一,合并框的位置是这两个框位置的均值,大小也是这两个框大小的均值。最后剩下少数几个人脸框,每个框都由一定数量的候选框合并而成,这个数值是一个非常重要的参数,决定了检测的人脸框是否是一个真实的人脸。本文设定一个阈值,如果合并框数大于这个阈值,则人脸当前位置是一个真实人脸,否则淘汰该候选框。图12b是合并后的结果。Inversely transform the square frames detected by two-level thick and thin in all scale images to the scale and position of the original input image, and obtain the candidate position and candidate size of the face in the original image. In general, a real face is often detected multiple times at different scales and adjacent positions, while the occurrence of false alarms is often relatively isolated. Figure 12a is an example. At this time, it is necessary to post-process the detection results and merge the face frames in adjacent positions. If the size difference and position difference of the two candidate face frames are very small, or the overlapping area of the two face frames is very large, you can Combine these two boxes into one, the position of the combined box is the mean of the positions of the two boxes, and the size is also the mean of the sizes of the two boxes. In the end, a few face frames are left, and each frame is formed by merging a certain number of candidate frames. This value is a very important parameter, which determines whether the detected face frame is a real face. This paper sets a threshold. If the number of merged frames is greater than this threshold, the current position of the face is a real face, otherwise the candidate frame is eliminated. Figure 12b is the combined result.
前面提到的人脸检测算法仅能检测正面人脸,能容忍的人脸深度变化、平面旋转角度都非常有限。另外,为获取一幅图像中的人脸,人脸检测算法需要对不同尺度的缩放图像进行大量搜索,尽管算法可以在数十毫秒内实现人脸的检测,但这个过程仍旧是非常耗时和耗计算量的。如果对实时输入的视频序列的每帧图像都进行人脸检测,则整个算法的计算量将非常惊人。人脸跟踪另外一个重要的目的就是实现某个人脸的持续跟踪,确认长时间跟踪的目标是同一个人脸,这样后续的人脸处理算法如人脸识别、表情识别等可以综合视频中多帧识别结果,大大提高识别算法的精度。The face detection algorithm mentioned above can only detect frontal faces, and the face depth changes and plane rotation angles that can be tolerated are very limited. In addition, in order to obtain the face in an image, the face detection algorithm needs to search a large number of zoomed images of different scales. Although the algorithm can detect the face within tens of milliseconds, this process is still very time-consuming and expensive. computationally consuming. If the face detection is performed on each frame of the video sequence input in real time, the calculation amount of the whole algorithm will be very alarming. Another important purpose of face tracking is to achieve continuous tracking of a certain face, to confirm that the target of long-term tracking is the same face, so that subsequent face processing algorithms such as face recognition, expression recognition, etc. can comprehensively recognize multiple frames in the video As a result, the accuracy of the recognition algorithm is greatly improved.
人脸的跟踪是在人脸检测的基础上实现。先对视频序列的输入图像进行人脸检测,为了降低程序CPU的占有率,可以每隔几帧检测一次人脸。检测到人脸后在后续两帧中对检测人脸进行跟踪和验证,留下验证通过的人脸中最大的一个,持续跟踪,处理人脸的各种姿态变换。考虑到人脸面部的肤色有着非常鲜明的特征,与头发、衣服、拍摄场景的差异都非常大,因此本发明也采用肤色特征实现人脸的跟踪,而肤色特征的获取通过颜色直方图特征实现。如图13,本文计算圆形区域内的颜色直方图,并利用第k-1帧的人脸坐标、直方图特征到第k帧图像中进行搜索,获取第k帧的人脸位置。Face tracking is realized on the basis of face detection. Face detection is first performed on the input image of the video sequence. In order to reduce the CPU usage of the program, the face can be detected every few frames. After the face is detected, the detected face is tracked and verified in the next two frames, and the largest face among the verified faces is left, which is continuously tracked and various pose changes of the face are processed. Considering that the skin color of the human face has very distinctive features, which are very different from hair, clothes, and shooting scenes, the present invention also uses the skin color feature to track the face, and the skin color feature is obtained through the color histogram feature. . As shown in Figure 13, this paper calculates the color histogram in the circular area, and uses the face coordinates and histogram features of the k-1th frame to search the image of the kth frame to obtain the face position of the kth frame.
本发明在进行人脸跟踪的时候,采用了基于Mean shift和直方图特征的物体跟踪算法,该算法正是采用直方图特征实现某个颜色目标的快速跟踪,算法的处理效率非常高,该算法与人脸检测结合起来,实现视频序列中人脸的持续检测与跟踪。The present invention adopts an object tracking algorithm based on Mean shift and histogram features when performing face tracking. This algorithm uses the histogram features to realize fast tracking of a certain color target. The processing efficiency of the algorithm is very high. Combined with face detection, it realizes the continuous detection and tracking of faces in video sequences.
本发明所述的方法,在计算人脸颜色直方图特征时,可将R、G、B每个颜色空间量化为8级,总的颜色空间量化为8×8×8级,这样每次计算的直方图特征为512维。人脸区域有三个参数描述(xcen,ycen,rad),分别表示人脸中心点的xy坐标、圆形人脸的半径,如图13a,当然在实际应用的时候也可以采用其他的空间量化级实现。当一幅新的图像输入时,跟踪算法根据前一帧的人脸位置大小以及直方图特征计算当前帧人脸的新位置,并更新人脸的半径,反映人脸大小的变换。The method of the present invention can quantize each color space of R, G, and B into 8 levels when calculating the face color histogram feature, and the total color space can be quantized into 8 × 8 × 8 levels, so that each calculation The histogram feature of is 512-dimensional. The face area is described by three parameters (x cen , y cen , rad), which respectively represent the xy coordinates of the center point of the face and the radius of the circular face, as shown in Figure 13a. Of course, other spaces can also be used in practical applications Quantitative level implementation. When a new image is input, the tracking algorithm calculates the new position of the face in the current frame based on the size of the face in the previous frame and the histogram features, and updates the radius of the face to reflect the transformation of the face size.
本发明在跟踪的过程中引入了长期直方图和短期直方图两个局部特征,其中长期直方图是前面数十帧直方图的均值,反映了人脸相当长时间内肤色的变化规律,而短期直方图则是前面数帧直方图的均值,反映的则是人脸短时间内肤色的变化。在搜索当前帧人脸的位置时,采用的匹配直方图模板等于长期直方图特征与短期直方图特征的均值,这样即使当前帧人脸否认姿态、光照、表情等变化剧烈,其肤色特征与直方图模板的差异也不会太大,采用Mean shift算法(运动跟踪算法)就可快速获取人脸的位置。The present invention introduces two local features of long-term histogram and short-term histogram in the process of tracking, wherein the long-term histogram is the mean value of the histogram of the previous dozens of frames, which reflects the changing law of the skin color of the face for a long time, while the short-term histogram The histogram is the average value of the histogram of the previous frames, which reflects the change of the skin color of the face in a short period of time. When searching for the position of the face in the current frame, the matching histogram template used is equal to the mean value of the long-term histogram feature and the short-term histogram feature, so that even if the face in the current frame denies the pose, illumination, expression, etc. The difference between the image templates will not be too large, and the position of the face can be quickly obtained by using the Mean shift algorithm (motion tracking algorithm).
为保证跟踪目标一定是人脸,并确保半径rad可准确描述人脸尺寸的变化,本发明在人脸跟踪时加入了跟踪区域的验证,即每隔数帧在跟踪区域进行正面人脸检测,此时人脸的位置与尺寸已经近似知道,因此不需要搜索整幅图像,只需要搜索少数位置和少数尺度即可。如果在跟踪区域检测到正面人脸,系统再根据检测人脸的大小和位置更新跟踪参数,包括人脸的中心、半径和直方图特征。另外,如果连续数百帧都跟踪到目标,但在跟踪区域都没有检测到正面人脸,此时可以认为目标不一定是人脸,停止跟踪。在后续帧中对输入图像进行穷搜索,重新检测人脸并跟踪。该方法可以用于人脸识别、表情识别、人脸合成等方法中去,以便进行实时检测和跟踪。In order to ensure that the tracking target must be a human face, and to ensure that the radius rad can accurately describe the change of the size of the human face, the present invention adds the verification of the tracking area during face tracking, that is, the frontal face detection is performed in the tracking area every few frames, At this time, the position and size of the face are approximately known, so there is no need to search the entire image, only a few positions and a few scales are needed. If a frontal face is detected in the tracking area, the system updates the tracking parameters according to the size and position of the detected face, including the center, radius and histogram features of the face. In addition, if the target is tracked for hundreds of consecutive frames, but no frontal face is detected in the tracking area, it can be considered that the target is not necessarily a human face at this time, and the tracking is stopped. Perform a poor search on the input image in subsequent frames, re-detect the face and track it. This method can be used in face recognition, expression recognition, face synthesis and other methods for real-time detection and tracking.
本发明的人脸检测算法可准确检测-20°到20°平面旋转人脸、-20°到20°左右深度旋转人脸,可以检测一定范围的抬头和低头人脸,可以检测不同表情人脸,戴与不戴眼镜对检测效果没有影响。图14是一组人脸检测结果,其中的图14a中每个人脸图像包括多个候选框和一个处理后的检测结果框,其中线141所示的框表示后处理后的检测结果,线142所示的框表示的为一个候选框,在本图中包含了多个候选框;图14b-d中的框是进行后处理后的输出结果,图14c-d中人脸左右两侧光照差异较大,但本发明在方法中采用了左右灰度归一化处理的方法,提高了检测算法抗光照干扰的能力,实现了这类人脸的准确检测,图15是一组跟踪结果。The face detection algorithm of the present invention can accurately detect faces rotated in a plane from -20° to 20°, and faces rotated in depth from -20° to 20°, and can detect faces with heads up and down in a certain range, and faces with different expressions , wearing or not wearing glasses has no effect on the detection effect. Fig. 14 is a group of face detection results, wherein each face image in Fig. 14a includes a plurality of candidate frames and a processed detection result frame, wherein the frame shown in
本发明上述的方法可由视频序列中人脸的实时检测与持续跟踪系统实现,参考图16的内容,该系统包括人脸检测装置1与人脸跟踪装置2,人脸检测装置1包括人脸处理单元11、微特征计算单元12及分类器单元14;所述的人脸处理单元11接收到图像信息,对接收到的图像进行缩放,穷举搜索缩放图像中的候选人脸窗口,计算窗口图像左右半边的均值和方差,然后将其传送给所述的微特征计算单元12,所述的特征计算单元12根据AdaBoost算法计算出微结构特征,并将其传送给所述的分类器单元14进行判决,分类器单元14进行判决后将其传送给人脸跟踪装置2;所述的人脸跟踪装置2包括物体跟踪单元21和跟踪区域验证单元22,物体跟踪单元21采用直方图特征进行计算,实现对图像跟踪,跟踪区域验证单元22对跟踪的图像进行区域检测,对跟踪的人脸进行验证。The above-mentioned method of the present invention can be realized by the real-time detection and continuous tracking system of human faces in video sequences. With reference to the content of FIG. 16, the system includes a human face detection device 1 and a human
在本发明所述的系统中,该人脸检测装置1的分类器单元14还包含粗细两级人脸检测单元13,接收微结构特征,进行粗细两级检测和后处理,确定人脸窗口,再进行跟踪。这种系统的构成框图可参见图17所示的内容。In the system of the present invention, the
本发明实现了实时输入的视频序列中人脸的检测与持续跟踪。该算法采用基于AdaBoost统计分层分类器的人脸检测方法,实现正面直立人脸的实时检测,并可检测不同场景不同表情的人脸,容忍一定范围和一定角度的姿态、旋转变化,并且采用基于Mean shift和直方图特征的人脸跟踪方法,实现了检测人脸的实时跟踪,跟踪算法速度块,CPU占有率低,且不受人脸姿态的影响,侧面、旋转人脸同样可以跟踪。The invention realizes the detection and continuous tracking of the human face in the real-time input video sequence. The algorithm adopts the face detection method based on the AdaBoost statistical hierarchical classifier to realize real-time detection of upright faces, and can detect faces with different expressions in different scenes. The face tracking method based on Mean shift and histogram features realizes real-time tracking of detected faces. The speed of the tracking algorithm is fast, the CPU occupancy rate is low, and it is not affected by the posture of the face. Sideways and rotated faces can also be tracked.
以上所述仅为本发明的较佳实施例而已,并不用以限制本发明,凡在本发明的精神和原则之内,所作的任何修改、等同替换等,均应包含在本发明的保护范围之内。The above descriptions are only preferred embodiments of the present invention, and are not intended to limit the present invention. Any modifications, equivalent replacements, etc. made within the spirit and principles of the present invention should be included in the protection scope of the present invention within.
Claims (14)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CNB2005101356688A CN100361138C (en) | 2005-12-31 | 2005-12-31 | Method and system for real-time detection and continuous tracking of human faces in video sequences |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CNB2005101356688A CN100361138C (en) | 2005-12-31 | 2005-12-31 | Method and system for real-time detection and continuous tracking of human faces in video sequences |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN1794264A CN1794264A (en) | 2006-06-28 |
| CN100361138C true CN100361138C (en) | 2008-01-09 |
Family
ID=36805689
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CNB2005101356688A Expired - Fee Related CN100361138C (en) | 2005-12-31 | 2005-12-31 | Method and system for real-time detection and continuous tracking of human faces in video sequences |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN100361138C (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111898581A (en) * | 2020-08-12 | 2020-11-06 | 成都佳华物链云科技有限公司 | Animal detection method, device, electronic equipment and readable storage medium |
Families Citing this family (53)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN100397411C (en) * | 2006-08-21 | 2008-06-25 | 北京中星微电子有限公司 | People face track display method and system for real-time robust |
| CN100426317C (en) * | 2006-09-27 | 2008-10-15 | 北京中星微电子有限公司 | Multiple attitude human face detection and track system and method |
| CN100426318C (en) * | 2006-09-28 | 2008-10-15 | 北京中星微电子有限公司 | AAM-based object location method |
| CN101178769B (en) * | 2007-12-10 | 2013-03-27 | 北京中星微电子有限公司 | Health protecting equipment and realization method thereof |
| CN101221620B (en) * | 2007-12-20 | 2011-04-06 | 北京中星微电子有限公司 | Human face tracing method |
| CN101499128B (en) * | 2008-01-30 | 2011-06-29 | 中国科学院自动化研究所 | 3D Face Action Detection and Tracking Method Based on Video Stream |
| CN101339608B (en) * | 2008-08-15 | 2011-10-12 | 北京中星微电子有限公司 | Object tracking method and system based on detection |
| CN101383001B (en) * | 2008-10-17 | 2010-06-02 | 中山大学 | A Fast and Accurate Frontal Face Discrimination Method |
| NO329897B1 (en) * | 2008-12-19 | 2011-01-24 | Tandberg Telecom As | Procedure for faster face detection |
| CN101447023B (en) * | 2008-12-23 | 2013-03-27 | 北京中星微电子有限公司 | Method and system for detecting human head |
| CN101456501B (en) * | 2008-12-30 | 2014-05-21 | 北京中星微电子有限公司 | Method and device for controlling elevator buttons |
| CN101577812B (en) * | 2009-03-06 | 2014-07-30 | 北京中星微电子有限公司 | Method and system for post monitoring |
| TWI401963B (en) * | 2009-06-25 | 2013-07-11 | Pixart Imaging Inc | Dynamic image compression method for face detection |
| CN101661554B (en) * | 2009-09-29 | 2012-02-01 | 哈尔滨工程大学 | Automatic identification method of frontal human body under long-distance video |
| CN101706721B (en) * | 2009-12-21 | 2012-11-28 | 汉王科技股份有限公司 | Face detection method simulating radar scanning |
| CN101794382B (en) * | 2010-03-12 | 2012-06-13 | 华中科技大学 | Method for counting passenger flow of buses in real time |
| US9082001B2 (en) | 2010-06-23 | 2015-07-14 | Nokia Technologies Oy | Method, apparatus and computer program product for tracking face portion |
| US9628755B2 (en) * | 2010-10-14 | 2017-04-18 | Microsoft Technology Licensing, Llc | Automatically tracking user movement in a video chat application |
| CN102004899B (en) * | 2010-11-03 | 2012-09-26 | 无锡中星微电子有限公司 | Human face identifying system and method |
| CN102004905B (en) * | 2010-11-18 | 2012-11-21 | 无锡中星微电子有限公司 | Human face authentication method and device |
| CN102004909A (en) * | 2010-11-30 | 2011-04-06 | 方正国际软件有限公司 | Method and system for processing identity information |
| CN102110399B (en) * | 2011-02-28 | 2016-08-24 | 北京中星微电子有限公司 | A kind of assist the method for explanation, device and system thereof |
| CN103020580B (en) * | 2011-09-23 | 2015-10-28 | 无锡中星微电子有限公司 | Fast face detecting method |
| CN102722698B (en) * | 2012-05-17 | 2014-03-12 | 上海中原电子技术工程有限公司 | Method and system for detecting and tracking multi-pose face |
| CN103218603B (en) * | 2013-04-03 | 2016-06-01 | 哈尔滨工业大学深圳研究生院 | A kind of face automatic marking method and system |
| CN103310466B (en) * | 2013-06-28 | 2016-02-17 | 安科智慧城市技术(中国)有限公司 | A kind of monotrack method and implement device thereof |
| CN104866805B (en) * | 2014-02-20 | 2020-09-11 | 腾讯科技(深圳)有限公司 | Method and device for real-time tracking of human face |
| CN104036237B (en) * | 2014-05-28 | 2017-10-10 | 中国人民解放军海军总医院 | The detection method of rotation face based on on-line prediction |
| SG10201504080WA (en) * | 2015-05-25 | 2016-12-29 | Trakomatic Pte Ltd | Method and System for Facial Recognition |
| CN106326817B (en) * | 2015-07-03 | 2021-08-03 | 佳能株式会社 | Method and apparatus for detecting object from image |
| CN106469289A (en) * | 2015-08-16 | 2017-03-01 | 联芯科技有限公司 | Facial image sex-screening method and system |
| CN206214373U (en) * | 2016-03-07 | 2017-06-06 | 维看公司 | Object detection from visual information to blind person, analysis and prompt system for providing |
| TW201743241A (en) | 2016-06-01 | 2017-12-16 | 原相科技股份有限公司 | Portable electronic device and operation method thereof |
| CN113378138A (en) * | 2016-06-15 | 2021-09-10 | 原相科技股份有限公司 | Portable electronic device |
| CN106339693A (en) * | 2016-09-12 | 2017-01-18 | 华中科技大学 | Positioning method of face characteristic point under natural condition |
| CN106407966B (en) * | 2016-11-28 | 2019-10-18 | 南京理工大学 | A face recognition method applied to attendance |
| CN106919903B (en) * | 2017-01-19 | 2019-12-17 | 中国科学院软件研究所 | A Robust Deep Learning-Based Method for Continuous Emotion Tracking |
| CN107292284B (en) * | 2017-07-14 | 2020-02-28 | 成都通甲优博科技有限责任公司 | Target re-detection method and device and unmanned aerial vehicle |
| CN107633208B (en) * | 2017-08-17 | 2018-12-18 | 平安科技(深圳)有限公司 | Electronic device, the method for face tracking and storage medium |
| CN108564037B (en) * | 2018-04-15 | 2021-06-08 | 南京明辉创鑫电子科技有限公司 | Salutation posture detection and correction method |
| CN109614841B (en) * | 2018-04-26 | 2023-04-18 | 杭州智诺科技股份有限公司 | Rapid face detection method in embedded system |
| CN109145771B (en) * | 2018-08-01 | 2020-11-20 | 武汉普利商用机器有限公司 | Face snapshot method and device |
| CN109271848B (en) * | 2018-08-01 | 2022-04-15 | 深圳市天阿智能科技有限责任公司 | Face detection method, face detection device and storage medium |
| CN109543534B (en) * | 2018-10-22 | 2020-09-01 | 中国科学院自动化研究所南京人工智能芯片创新研究院 | Method and device for re-detecting lost target in target tracking |
| CN109670451A (en) * | 2018-12-20 | 2019-04-23 | 天津天地伟业信息系统集成有限公司 | Automatic face recognition tracking |
| CN109946229B (en) * | 2019-02-25 | 2024-07-16 | 南京文易特电子科技有限公司 | Intelligent digital double-pull-wire detection system and detection method for cigarette strip package |
| CN110288632A (en) * | 2019-05-15 | 2019-09-27 | 北京旷视科技有限公司 | A kind of image processing method, device, terminal and storage medium |
| CN110287778B (en) * | 2019-05-15 | 2021-09-10 | 北京旷视科技有限公司 | Image processing method and device, terminal and storage medium |
| CN112101063A (en) * | 2019-06-17 | 2020-12-18 | 福建天晴数码有限公司 | Skew face detection method and computer-readable storage medium |
| CN110555867B (en) * | 2019-09-05 | 2023-07-07 | 杭州智爱时刻科技有限公司 | Multi-target object tracking method integrating object capturing and identifying technology |
| CN112749603B (en) * | 2019-10-31 | 2024-09-17 | 上海商汤智能科技有限公司 | Living body detection method, living body detection device, electronic equipment and storage medium |
| CN112153334B (en) * | 2020-09-15 | 2023-02-21 | 公安部第三研究所 | Intelligent video box device for security management and corresponding intelligent video analysis method |
| CN117437678A (en) * | 2023-11-01 | 2024-01-23 | 烟台持久钟表有限公司 | Front face duration statistics method, system, device and storage medium |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6301370B1 (en) * | 1998-04-13 | 2001-10-09 | Eyematic Interfaces, Inc. | Face recognition from video images |
| US20040109584A1 (en) * | 2002-09-18 | 2004-06-10 | Canon Kabushiki Kaisha | Method for tracking facial features in a video sequence |
| WO2004051551A1 (en) * | 2002-11-29 | 2004-06-17 | Sony United Kingdom Limited | Face detection and tracking |
| WO2005096213A1 (en) * | 2004-03-05 | 2005-10-13 | Thomson Licensing | Face recognition system and method |
-
2005
- 2005-12-31 CN CNB2005101356688A patent/CN100361138C/en not_active Expired - Fee Related
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6301370B1 (en) * | 1998-04-13 | 2001-10-09 | Eyematic Interfaces, Inc. | Face recognition from video images |
| US20040109584A1 (en) * | 2002-09-18 | 2004-06-10 | Canon Kabushiki Kaisha | Method for tracking facial features in a video sequence |
| WO2004051551A1 (en) * | 2002-11-29 | 2004-06-17 | Sony United Kingdom Limited | Face detection and tracking |
| WO2005096213A1 (en) * | 2004-03-05 | 2005-10-13 | Thomson Licensing | Face recognition system and method |
Non-Patent Citations (3)
| Title |
|---|
| 一种新的Adaboost快速训练算法. 王海川,张立明.复旦学报(自然科学版),第43卷第1期. 2004 * |
| 动态人脸表情识别技术研究. 应伟,第27-28页,第36-37页,中国优秀博硕士学位论文全文数据库. 2005 * |
| 采用Adaboost算法进行面部表情识别. 杨国亮,王志良,任金霞.计算机应用,第25卷第4期. 2005 * |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111898581A (en) * | 2020-08-12 | 2020-11-06 | 成都佳华物链云科技有限公司 | Animal detection method, device, electronic equipment and readable storage medium |
| CN111898581B (en) * | 2020-08-12 | 2024-05-17 | 成都佳华物链云科技有限公司 | Animal detection method, apparatus, electronic device, and readable storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| CN1794264A (en) | 2006-06-28 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN100361138C (en) | Method and system for real-time detection and continuous tracking of human faces in video sequences | |
| Wu et al. | Simultaneous object detection and segmentation by boosting local shape feature based classifier | |
| Tsai et al. | Vehicle detection using normalized color and edge map | |
| CN108520226B (en) | A pedestrian re-identification method based on body decomposition and saliency detection | |
| CN101339607B (en) | Human face recognition method and system, human face recognition model training method and system | |
| CN101872477B (en) | Method and device for detecting object in image and system containing device | |
| Sahbi et al. | A Hierarchy of Support Vector Machines for Pattern Detection. | |
| CN102214309B (en) | Special human body recognition method based on head and shoulder model | |
| CN103632132A (en) | Face detection and recognition method based on skin color segmentation and template matching | |
| CN104091147A (en) | Near infrared eye positioning and eye state identification method | |
| JP2014093023A (en) | Object detection device, object detection method and program | |
| CN101996308A (en) | Human face identification method and system and human face model training method and system | |
| Liu et al. | Multi-type road marking recognition using adaboost detection and extreme learning machine classification | |
| CN100440246C (en) | A Face Feature Point Locating Method | |
| Gürel et al. | Design of a face recognition system | |
| Vij et al. | A survey on various face detecting and tracking techniques in video sequences | |
| US7548637B2 (en) | Method for detecting objects in an image using pair-wise pixel discriminative features | |
| CN104200202B (en) | A kind of upper half of human body detection method based on cumulative perceptron | |
| CN110458064A (en) | Combining data-driven and knowledge-driven low-altitude target detection and recognition methods | |
| Efraty et al. | Facial component-landmark detection | |
| CN104517300A (en) | Vision judgment tracking method based on statistical characteristic | |
| Yang et al. | Vehicle detection under tough conditions using prioritized feature extraction with shadow recognition | |
| Curran et al. | The use of neural networks in real-time face detection | |
| Yu et al. | Research on face recognition method based on deep learning | |
| Intan et al. | Facial recognition using multi edge detection and distance measure |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| C06 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| C14 | Grant of patent or utility model | ||
| GR01 | Patent grant | ||
| TR01 | Transfer of patent right | ||
| TR01 | Transfer of patent right |
Effective date of registration: 20180410 Address after: 100191 Xueyuan Road, Haidian District, Haidian District, Beijing, No. 607, No. six Patentee after: BEIJING VIMICRO ARTIFICIAL INTELLIGENCE CHIP TECHNOLOGY CO.,LTD. Address before: 100083, Haidian District, Xueyuan Road, Beijing No. 35, Nanjing Ning building, 15 Floor Patentee before: VIMICRO Corp. |
|
| CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20080109 |