[go: up one dir, main page]

WO2014190494A1 - Method and device for facial recognition - Google Patents

Method and device for facial recognition Download PDF

Info

Publication number
WO2014190494A1
WO2014190494A1 PCT/CN2013/076343 CN2013076343W WO2014190494A1 WO 2014190494 A1 WO2014190494 A1 WO 2014190494A1 CN 2013076343 W CN2013076343 W CN 2013076343W WO 2014190494 A1 WO2014190494 A1 WO 2014190494A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
names
reference image
person
obtaining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2013/076343
Other languages
French (fr)
Inventor
Yanfeng Zhang
Xiaojun Ma
Yuntao Shi
Jun Xu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Thomson Licensing SAS
Original Assignee
Thomson Licensing SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing SAS filed Critical Thomson Licensing SAS
Priority to PCT/CN2013/076343 priority Critical patent/WO2014190494A1/en
Publication of WO2014190494A1 publication Critical patent/WO2014190494A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • G06F16/784Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content the detected or recognised objects being people

Definitions

  • the present invention relates to data processing, and more particularly relates to a method and device for facial recognition.
  • a facial recognition system is a computer application for automatically identifying or verifying a person from a digital image or a video frame from a video source.
  • One of the ways to do this is by comparing selected facial features from the image with a facial database.
  • the core technology for facial recognition is Local Feature Analysis (LFA) algorithm.
  • LFA Local Feature Analysis
  • Facial recognition systems use various LFA techniques to encode, match and recognize faces.
  • To recognize a face in a video a traditional method firstly locates human faces in a video frame, and then recognizes those faces by matching them against the database of known reference faces.
  • a database with a huge number of reference faces should be built in advance.
  • a method for recognizing a person in a video comprises steps of obtaining an image frame from the video containing the person; and recognizing the person by matching the image frame in a reference image database, wherein the reference image database consists of reference images associated with persons appearing in the video.
  • the reference image database stores the relationship between names and reference images, the method further comprising presenting a name that corresponding to the matched reference image in the reference image database. Further, the reference image database stores the relationship between names and reference images, the method further comprising searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and presenting the relevant information.
  • the method further comprising obtaining a list of names for the persons appearing in the video; obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
  • the step of obtaining the list of names further comprising obtaining name of the video; obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
  • a device for recognizing a person in a video comprising a reference image database module for obtaining a reference image database consisted of reference images associated with persons appearing in the video; and a facial recognition module for recognizing the person contained in an image frame from the video by matching the image frame with the reference images in the reference image database.
  • the reference image database stores the relationship between names and reference images
  • the device further comprising a presenting module for presenting a name that corresponding to the matched reference image in the reference image database.
  • the reference image database stores the relationship between names and reference images
  • the device further comprising a searching module for searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and a presenting module for presenting the relevant information.
  • the device further comprises a list fetching module for obtaining a list of names for the persons appearing in the video; wherein the reference image database module is further used for obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
  • the list fetching module is further used for obtaining name of the video; obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
  • Fig. 1 is a diagram showing a system architecture for facial recognition according to an embodiment of present invention
  • Fig. 2 is a diagram showing the key frame extraction according to the embodiment of present invention.
  • Fig. 3 is a flow chart showing a method for obtaining a cast list for a video according to the embodiment of present invention
  • Fig. 4 is a flow chart showing steps for face recognition according to the embodiment of the present invention
  • Fig. 5 is a diagram showing a web page for presenting information to the user according to the embodiment of the present invention.
  • the present invention provides a method for effective facial recognition.
  • the method can be used, for example, for effectively recognizing actors in a movie or a TV episode and providing their relevant information to users.
  • the method searches and retrieves cast list or starring list from movie/TV database website on the Internet, such as lmdb.com, allrovi.com, findanyfilm.com, TV.com, etc.
  • the actor names in the cast list are used to build a temporary CLFI (cast list face image) database.
  • Each temporary CLFI database corresponds to a specific movie or TV episode, and is built temporarily just for currently watched movie or TV episode.
  • Fig.1 shows a system architecture for facial recognition in terms of function according to an embodiment of present invention. It shall note that a person skilled in the art may contemplate a different system architecture for implementing the principle of the present invention by combining functions among modules as shown in the Fig. 1 .
  • the system is conceptually divided into 3 blocks, i.e. user side, server side and the Internet. It shall note that a person skilled in the art may divide the system in a different way while implementing the principle of the present invention.
  • the server side is incorporated into the block of the Internet.
  • Actor image obtaining module 101 is located on user's device, such as a computer, a television, a PC tablet, or a smartphone, etc., and used to obtain an image containing the actor that the user wants to know.
  • the module 101 concerns how to capture the actor image from a watching video, e.g. a movie or a TV episode. The following three methods can be used to capture the actor image for different application situations.
  • Fig. 2 is a diagram showing the key frame extraction according to the embodiment of present invention.
  • the first device is a TV set with the connection from air broadcast, satellite, fiber optic, cable or IP connection from the broadcaster.
  • the second device is equipped with an input- friendly component, such as touch screen.
  • This type of device includes laptop, tablet, PDA, smart phone etc.
  • a command message e.g. a capture command
  • the second device extracts image frame from the video clips and presents them on its screen. The user selects an appropriate image frame containing the actor to show it on the screen of the second device.
  • a camera such as a smart phone, a PC tablet, etc.
  • a smart device such as a computer, a PC tablet, or a smart phone
  • he can pause the video and instruct the device to copy current screen as the actor image.
  • the device can divide the image into sub images, each of which contains one person, and present them for the user to select. There are other ways for doing the selection of the actor. For example, the user can designate the targeting actor by drawing a cycle or a rectangle on the actor's face, or by just clicking on his face. Then the designated actor face image is sent to the server.
  • Modules of cast list fetching 102, face database building 103, facial recognition 104 and actor information searching and presenting 105 are located in the server side, for example in a server computer of service provider. It shall note that the server can reside in user's home or at the remote, e.g. service provider's premise.
  • Cast list fetching module 102 is configured to obtain a cast list of a movie or a TV episode from relevant movie/TV database websites on the Internet.
  • movie/TV database websites on the Internet for example, imdb.com, allrovi.com, findanyfilm.com and tv.com etc., usually provide an online database of information related to movies, television shows, actors, production crew personnel, video games and fictional characters featured in visual entertainment media.
  • Fig. 3 is a flow chart showing a method for obtaining a cast list for a video, e.g. a movie or a TV episode.
  • cast list fetching module 102 in the server gets information about the video being watched.
  • the information includes, but is not limited to, the name and release year of a movie, the name, the season number and the episode number for a TV episode. Normally, the information can be obtained from EPG (Electronic Program Guide). In an example, only the name of the video is used.
  • EPG Electronic Program Guide
  • the information is used to search in the first site of the movie/TV website list.
  • the website list includes websites capable of providing online information related to movies, television shows, actors, production crew personnel, video games and fictional characters featured in the video.
  • the website list includes imdb.com, allrovi.com, findanyfilm.com, tv.com, etc.
  • the step 303 it checks whether or not there is entry in the search result for the video. If no, in the step 304 it will continue to search the next website in the list until it reaches the end of the list. If yes, in the step 307 it will use one of keywords including "cast", "starring” and "actor” etc. to search for the cast list.
  • it checks if it succeeds in obtaining the cast list. If not, it will determine that no result is found and prompt such information to the user in the steps 305 and 306. If yes, in the step 309 it will sends the cast list to the face database building module.
  • Face database building module 103 is configured to use the actor names in the cast list to search for the actor's images in image searching engines, such as Google/Yahoo image search, or picture sharing website, for example, Flickr, and then build a temporary reference image database by using the searched images as reference images.
  • the database has at least two data fields, one data field contains the actor names and the other data field contains reference images for the actor. It shall note that one or more reference images correspond to one actor name.
  • the temporary reference image database is stored in the server. When a new cast list becomes available (e.g. proceeding to next episode), the server needs to update the temporary reference image database. The server firstly searches in the temporary reference image database with the actor names to find those names in the new cast list that don't have entry in the temporary reference image database. And then the server obtains reference images for those names and updates the temporary reference image database.
  • Facial recognizing module 104 is configured to recognize the actor in an image provided by the actor image obtaining module 101 by matching between the actor image provided by the actor image obtaining module 101 and the reference images in the temporary image database.
  • Facial recognition is a recognition process that analyzes facial characteristics, and then uses the characteristics to find similar or same faces in the database. There already exits multiple facial recognition algorithms, such as Eigenface, Fishferface and Elastic Graph Matching, Support Vector Machine, and Neural Networks, etc.
  • Fig. 4 is a flow chart showing steps for face recognition according to the embodiment of the present invention.
  • face detection and face alignment are the basis for feature extraction.
  • the actor image captured or taken from currently watched video is passed to face detection, which will go through the color model, skin segmentation to extract any possible face region from background scenes. Then the face alignment crops and resizes face areas to make sure they have the same sizes and scales. Face alignment also removes the noise and illumination interference, remedies affine transformation etc.
  • feature extraction is performed on the aligned face to provide effective information - face features that are useful for distinguishing between faces of different persons.
  • the lower part in the rectangle in the Fig. 4 is the feature extraction preprocess for reference images in the temporary reference image database. The preprocess procedure deals with all reference images in the temporary reference image database to extract their features and to store these features.
  • the preprocess uses the same face detection, face alignment and feature extraction as the above operation on the actor image.
  • the extracted actor image features are then compared with the features of the reference images for recognizing the actor.
  • Euclidean distance between the two groups of features is calculated. If the Euclidean distance is within a similarity threshold, it can be determined that the reference image matches with the actor image.
  • the value of similarity threshold depends on the actual feature extraction algorithm and system requirement.
  • the output of the facial recognition procedure is the actor name.
  • the actor information searching and presenting module 105 is configured to use the identified actor name to search in the movie/TV database websites on the Internet again to retrieve relevant information about the identified actor. It shall note that Internet search engine, such as Google, Bing, etc., can also be used to retrieve the relevant information about the identified actor.
  • the relevant information includes biography, photos including the actor, video clips including the actor, awards and nominations etc.
  • the name and the relevant information are organized and presented to the user.
  • Fig. 5 is a diagram showing a web page for presenting information to the user.
  • the functions realized by cast listing fetching module, face database building module and the function of feature extraction for reference images in the face database can be executed before actor image is provided by the actor image obtaining module 101 . That is to say, when user starts to see a video, the system automatically executes the above three functions.
  • the method can be used in sports event, e.g. golf, boxing, etc.
  • the method can further be used for a video as long as there is a list of names associated with the video, i.e. a list of names for persons who appear in the video.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Library & Information Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Processing Or Creating Images (AREA)

Abstract

It is provided a method for recognizing a person in a video, wherein the method comprises steps of obtaining an image frame from the video containing the person; and recognizing the person by matching the image frame in a reference image database, wherein the reference image database consists of reference images associated with persons appearing in the video.

Description

METHOD AND DEVICE FOR FACIAL RECOGNITION
TECHNICAL FIELD
The present invention relates to data processing, and more particularly relates to a method and device for facial recognition.
BACKGROUND
A facial recognition system is a computer application for automatically identifying or verifying a person from a digital image or a video frame from a video source. One of the ways to do this is by comparing selected facial features from the image with a facial database. The core technology for facial recognition is Local Feature Analysis (LFA) algorithm. Facial recognition systems use various LFA techniques to encode, match and recognize faces. To recognize a face in a video, a traditional method firstly locates human faces in a video frame, and then recognizes those faces by matching them against the database of known reference faces. Using the traditional method to recognize actors in movies or TV series, a database with a huge number of reference faces should be built in advance. SUMMARY
According to an aspect of the present invention, it is provided a method for recognizing a person in a video, wherein the method comprises steps of obtaining an image frame from the video containing the person; and recognizing the person by matching the image frame in a reference image database, wherein the reference image database consists of reference images associated with persons appearing in the video.
Further, the reference image database stores the relationship between names and reference images, the method further comprising presenting a name that corresponding to the matched reference image in the reference image database. Further, the reference image database stores the relationship between names and reference images, the method further comprising searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and presenting the relevant information.
Further, the method further comprising obtaining a list of names for the persons appearing in the video; obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
Further, the step of obtaining the list of names further comprising obtaining name of the video; obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
According to an aspect of the present invention, it is provided a device for recognizing a person in a video, wherein the device comprising a reference image database module for obtaining a reference image database consisted of reference images associated with persons appearing in the video; and a facial recognition module for recognizing the person contained in an image frame from the video by matching the image frame with the reference images in the reference image database.
Further, the reference image database stores the relationship between names and reference images, the device further comprising a presenting module for presenting a name that corresponding to the matched reference image in the reference image database.
Further, the reference image database stores the relationship between names and reference images, the device further comprising a searching module for searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and a presenting module for presenting the relevant information.
Further, the device further comprises a list fetching module for obtaining a list of names for the persons appearing in the video; wherein the reference image database module is further used for obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
Further, the list fetching module is further used for obtaining name of the video; obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
It is to be understood that more aspects and advantages of the invention will be found in the following detailed description of the present invention.
BRIEF DESCRIPTION OF THE DRAWINGS
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, will be used to illustrate an embodiment of the invention, as explained by the description. The invention is not limited to the embodiment.
In the drawings:
Fig. 1 is a diagram showing a system architecture for facial recognition according to an embodiment of present invention;
Fig. 2 is a diagram showing the key frame extraction according to the embodiment of present invention;
Fig. 3 is a flow chart showing a method for obtaining a cast list for a video according to the embodiment of present invention;
Fig. 4 is a flow chart showing steps for face recognition according to the embodiment of the present invention; Fig. 5 is a diagram showing a web page for presenting information to the user according to the embodiment of the present invention.
DETAILED DESCRI PTION The embodiment of the present invention will now be described in detail in conjunction with the drawings. In the following description, some detailed descriptions of known functions and configurations may be omitted for clarity and conciseness.
The present invention provides a method for effective facial recognition. The method can be used, for example, for effectively recognizing actors in a movie or a TV episode and providing their relevant information to users. With the movie name or TV series name and other related information, such as release year of the movie, season number and episode number of a TV episode, the method searches and retrieves cast list or starring list from movie/TV database website on the Internet, such as lmdb.com, allrovi.com, findanyfilm.com, TV.com, etc. Then the actor names in the cast list are used to build a temporary CLFI (cast list face image) database. Each temporary CLFI database corresponds to a specific movie or TV episode, and is built temporarily just for currently watched movie or TV episode. In a movie or TV episode's CLFI database, only images relating to actors in the cast list are contained for facial recognition and matching. Such images are obtained by searching in image search website (such as Google image search, yahoo image search) or picture sharing website (for example Flickr) with the actor names in the cast list. With the CLFI database of a movie or a TV episode, a facial recognition algorithm is applied to find who the actor is. Besides, after getting the actor's name, it is optional to retrieve and provide relevant information about the actor to users. Such information may include, but is not limited to, biography, awards or nomination, filmography, etc. The relevant information is retrieved by searching on the Internet. With the above method, the calculation requirement for facial recognition is minimized by using a temporary database, which is automatically built and contains reference images relevant to the actors in the cast list.
Fig.1 shows a system architecture for facial recognition in terms of function according to an embodiment of present invention. It shall note that a person skilled in the art may contemplate a different system architecture for implementing the principle of the present invention by combining functions among modules as shown in the Fig. 1 .
The system is conceptually divided into 3 blocks, i.e. user side, server side and the Internet. It shall note that a person skilled in the art may divide the system in a different way while implementing the principle of the present invention. For example, the server side is incorporated into the block of the Internet.
Actor image obtaining module 101 is located on user's device, such as a computer, a television, a PC tablet, or a smartphone, etc., and used to obtain an image containing the actor that the user wants to know. The module 101 concerns how to capture the actor image from a watching video, e.g. a movie or a TV episode. The following three methods can be used to capture the actor image for different application situations.
1 . Key frame extraction:
Fig. 2 is a diagram showing the key frame extraction according to the embodiment of present invention. As illustrated in fig.2, the first device is a TV set with the connection from air broadcast, satellite, fiber optic, cable or IP connection from the broadcaster. The second device is equipped with an input- friendly component, such as touch screen. This type of device includes laptop, tablet, PDA, smart phone etc. There is a communication channel between the first device and the second device to provide data exchange. When a user finds an actor he prefers to know in the video, he can send a command message (e.g. a capture command) via the second device to the first device, and then the first device will send back a short clip of the video. The second device extracts image frame from the video clips and presents them on its screen. The user selects an appropriate image frame containing the actor to show it on the screen of the second device.
2. Picture taking
While user is watching a movie or a TV episode, he uses a device with a camera, such as a smart phone, a PC tablet, etc., to take a picture that contains the actor he prefers to know. Besides, he can select to record a video clip with the device, and choose an image containing the actor from key frames extracted by the device.
3. Pausing a video and copying the screen.
While the user is watching a movie or a TV episode on a smart device, such as a computer, a PC tablet, or a smart phone, he can pause the video and instruct the device to copy current screen as the actor image.
If the captured image containing more than one person, the device can divide the image into sub images, each of which contains one person, and present them for the user to select. There are other ways for doing the selection of the actor. For example, the user can designate the targeting actor by drawing a cycle or a rectangle on the actor's face, or by just clicking on his face. Then the designated actor face image is sent to the server.
Modules of cast list fetching 102, face database building 103, facial recognition 104 and actor information searching and presenting 105 are located in the server side, for example in a server computer of service provider. It shall note that the server can reside in user's home or at the remote, e.g. service provider's premise.
Cast list fetching module 102 is configured to obtain a cast list of a movie or a TV episode from relevant movie/TV database websites on the Internet. Such movie/TV database websites on the Internet, for example, imdb.com, allrovi.com, findanyfilm.com and tv.com etc., usually provide an online database of information related to movies, television shows, actors, production crew personnel, video games and fictional characters featured in visual entertainment media.
Fig. 3 is a flow chart showing a method for obtaining a cast list for a video, e.g. a movie or a TV episode. In the step 301 , cast list fetching module 102 in the server gets information about the video being watched. The information includes, but is not limited to, the name and release year of a movie, the name, the season number and the episode number for a TV episode. Normally, the information can be obtained from EPG (Electronic Program Guide). In an example, only the name of the video is used. In the step 302, the information is used to search in the first site of the movie/TV website list. The website list includes websites capable of providing online information related to movies, television shows, actors, production crew personnel, video games and fictional characters featured in the video. In the embodiment, the website list includes imdb.com, allrovi.com, findanyfilm.com, tv.com, etc. In the step 303, it checks whether or not there is entry in the search result for the video. If no, in the step 304 it will continue to search the next website in the list until it reaches the end of the list. If yes, in the step 307 it will use one of keywords including "cast", "starring" and "actor" etc. to search for the cast list. In the step 308, it checks if it succeeds in obtaining the cast list. If not, it will determine that no result is found and prompt such information to the user in the steps 305 and 306. If yes, in the step 309 it will sends the cast list to the face database building module.
Face database building module 103 is configured to use the actor names in the cast list to search for the actor's images in image searching engines, such as Google/Yahoo image search, or picture sharing website, for example, Flickr, and then build a temporary reference image database by using the searched images as reference images. The database has at least two data fields, one data field contains the actor names and the other data field contains reference images for the actor. It shall note that one or more reference images correspond to one actor name. Besides, the temporary reference image database is stored in the server. When a new cast list becomes available (e.g. proceeding to next episode), the server needs to update the temporary reference image database. The server firstly searches in the temporary reference image database with the actor names to find those names in the new cast list that don't have entry in the temporary reference image database. And then the server obtains reference images for those names and updates the temporary reference image database.
Facial recognizing module 104 is configured to recognize the actor in an image provided by the actor image obtaining module 101 by matching between the actor image provided by the actor image obtaining module 101 and the reference images in the temporary image database.
With the actor image retrieved or taken from the currently watched video and with the temporary reference image database for the currently watched video, the facial recognizing module 104 is executed to find the identity of the actor. Facial recognition is a recognition process that analyzes facial characteristics, and then uses the characteristics to find similar or same faces in the database. There already exits multiple facial recognition algorithms, such as Eigenface, Fishferface and Elastic Graph Matching, Support Vector Machine, and Neural Networks, etc. Fig. 4 is a flow chart showing steps for face recognition according to the embodiment of the present invention.
In the Fig. 4, face detection and face alignment are the basis for feature extraction. The actor image captured or taken from currently watched video is passed to face detection, which will go through the color model, skin segmentation to extract any possible face region from background scenes. Then the face alignment crops and resizes face areas to make sure they have the same sizes and scales. Face alignment also removes the noise and illumination interference, remedies affine transformation etc. After face alignment, feature extraction is performed on the aligned face to provide effective information - face features that are useful for distinguishing between faces of different persons. The lower part in the rectangle in the Fig. 4 is the feature extraction preprocess for reference images in the temporary reference image database. The preprocess procedure deals with all reference images in the temporary reference image database to extract their features and to store these features. The preprocess uses the same face detection, face alignment and feature extraction as the above operation on the actor image. The extracted actor image features are then compared with the features of the reference images for recognizing the actor. To measure the similarity degree between the features of the actor image and the features of one reference image, Euclidean distance between the two groups of features is calculated. If the Euclidean distance is within a similarity threshold, it can be determined that the reference image matches with the actor image. The value of similarity threshold depends on the actual feature extraction algorithm and system requirement. The output of the facial recognition procedure is the actor name.
The actor information searching and presenting module 105 is configured to use the identified actor name to search in the movie/TV database websites on the Internet again to retrieve relevant information about the identified actor. It shall note that Internet search engine, such as Google, Bing, etc., can also be used to retrieve the relevant information about the identified actor. The relevant information includes biography, photos including the actor, video clips including the actor, awards and nominations etc. The name and the relevant information are organized and presented to the user. Fig. 5 is a diagram showing a web page for presenting information to the user.
In order to decrease response time for facial recognition and reduce the requirement on the calculation power of the server, the functions realized by cast listing fetching module, face database building module and the function of feature extraction for reference images in the face database can be executed before actor image is provided by the actor image obtaining module 101 . That is to say, when user starts to see a video, the system automatically executes the above three functions.
According to a variant, the method can be used in sports event, e.g. golf, boxing, etc. Besides, the method can further be used for a video as long as there is a list of names associated with the video, i.e. a list of names for persons who appear in the video.
According to a variant, if more than one person is included in an image frame, all persons will be identified and relevant information for all persons is provided and presented to user for selection of one targeting person.
A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. For example, elements of different implementations may be combined, supplemented, modified, or removed to produce other implementations. Additionally, one of ordinary skill will understand that other structures and processes may be substituted for those disclosed and the resulting implementations will perform at least substantially the same function(s), in at least substantially the same way(s), to achieve at least substantially the same result(s) as the implementations disclosed. Accordingly, these and other implementations are contemplated by this application and are within the scope of the invention as defined by the appended claims.

Claims

1 . A method for recognizing a person in a video, wherein the method comprsing steps of
obtaining an image frame from the video containing the person; and
recognizing the person by matching the image frame in a reference image database, wherein the reference image database consists of reference images associated with persons appearing in the video.
2. The method of the claim 1 , wherein the reference image database stores the relationship between names and reference images, the method further comprising presenting a name that corresponding to the matched reference image in the reference image database.
3. The method of the claim 1 , wherein the reference image database stores the relationship between names and reference images, the method further comprising searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and
presenting the relevant information.
4. The method of the claim 1 , wherein the method further comprising
obtaining a list of names for the persons appearing in the video;
obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
5. The method of the claim 4, wherein the step of obtaining the list of names further comprising obtaining name of the video;
obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
6. A device for recognizing a person in a video, wherein the device comprising a reference image database module for obtaining a reference image database consisted of reference images associated with persons appearing in the video; and a facial recognition module for recognizing the person contained in an image frame from the video by matching the image frame with the reference images in the reference image database.
7. The device of the claim 6, wherein the reference image database stores the relationship between names and reference images, the device further comprising a presenting module for presenting a name that corresponding to the matched reference image in the reference image database.
8. The device of the claim 6, wherein the reference image database stores the relationship between names and reference images, the device further comprising a searching module for searching for relevant information of the person by using name that corresponding to the matched reference image in the reference image database, wherein the relevant information comprises at least one of the following: picture having the person, video clip having the person, biography; and
a presenting module for presenting the relevant information.
9. The device of the claim 6, wherein the device further comprising
a list fetching module for obtaining a list of names for the persons appearing in the video;
wherein the reference image database module is further used for obtaining images corresponding to names in the list as the reference images from at least one information source providing correspondence relationship between persons' names and persons' images.
10. The device of the claim 9, wherein the list fetching module is further used for obtaining name of the video; obtaining the list by using the name to search in at least one information source providing correspondence relationship between video names and lists of names for persons appearing in the videos.
PCT/CN2013/076343 2013-05-28 2013-05-28 Method and device for facial recognition Ceased WO2014190494A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2013/076343 WO2014190494A1 (en) 2013-05-28 2013-05-28 Method and device for facial recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2013/076343 WO2014190494A1 (en) 2013-05-28 2013-05-28 Method and device for facial recognition

Publications (1)

Publication Number Publication Date
WO2014190494A1 true WO2014190494A1 (en) 2014-12-04

Family

ID=51987855

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/076343 Ceased WO2014190494A1 (en) 2013-05-28 2013-05-28 Method and device for facial recognition

Country Status (1)

Country Link
WO (1) WO2014190494A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180225377A1 (en) * 2015-06-17 2018-08-09 Baidu Online Network Technology (Beijing) Co., Ltd. Method, server and terminal for acquiring information and method and apparatus for constructing database
CN109377519A (en) * 2018-09-29 2019-02-22 佳都新太科技股份有限公司 Target tracking method, device, target tracking equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080175509A1 (en) * 2007-01-24 2008-07-24 General Electric Company System and method for reconstructing restored facial images from video
US20100149305A1 (en) * 2008-12-15 2010-06-17 Tandberg Telecom As Device and method for automatic participant identification in a recorded multimedia stream
WO2012122069A2 (en) * 2011-03-04 2012-09-13 Microsoft Corporation Aggregated facial tracking in video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080175509A1 (en) * 2007-01-24 2008-07-24 General Electric Company System and method for reconstructing restored facial images from video
US20100149305A1 (en) * 2008-12-15 2010-06-17 Tandberg Telecom As Device and method for automatic participant identification in a recorded multimedia stream
WO2012122069A2 (en) * 2011-03-04 2012-09-13 Microsoft Corporation Aggregated facial tracking in video

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180225377A1 (en) * 2015-06-17 2018-08-09 Baidu Online Network Technology (Beijing) Co., Ltd. Method, server and terminal for acquiring information and method and apparatus for constructing database
CN109377519A (en) * 2018-09-29 2019-02-22 佳都新太科技股份有限公司 Target tracking method, device, target tracking equipment and storage medium

Similar Documents

Publication Publication Date Title
US10333767B2 (en) Methods, systems, and media for media transmission and management
US10271098B2 (en) Methods for identifying video segments and displaying contextually targeted content on a connected television
US9176987B1 (en) Automatic face annotation method and system
JP6408551B2 (en) Method, system, and recording medium for providing portable content corresponding to media content
JP5358083B2 (en) Person image search device and image search device
US9110988B1 (en) Methods, systems, and media for aggregating and presenting multiple videos of an event
US9668002B1 (en) Identification of live streaming content
CN102868934B (en) Video object information search method and device based on intelligent television
KR101089287B1 (en) Automatic Face Recognition Apparatus and Method Based on Multi-face Feature Information Fusion
CN110119711A (en) A kind of method, apparatus and electronic equipment obtaining video data personage segment
US20150186511A1 (en) Methods, systems, and media for generating search results based on contextual information
WO2011090540A2 (en) Method for identifying video segments and displaying contextually targeted content on a connected television
CN103581705A (en) Method and system for recognizing video program
CN107480236A (en) A kind of information query method, device, equipment and medium
CN103686344A (en) Enhanced video system and method
KR20150083355A (en) Augmented media service providing method, apparatus thereof, and system thereof
CN106415546A (en) Systems and methods for locally detecting consumed video content
KR102298066B1 (en) Method for providing image contents and image contents providing apparatus
CN114339360B (en) Video processing method, related device and equipment
CN106528800B (en) A kind of image generating method and device based on real scene
WO2014190494A1 (en) Method and device for facial recognition
CN114329049B (en) Video search method, device, computer equipment and storage medium
CN111432279A (en) Method and device for controlling smart television and smart television
KR20150023492A (en) Synchronized movie summary
JP6192224B2 (en) Attention keyword information extraction device and program thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13885720

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13885720

Country of ref document: EP

Kind code of ref document: A1