US9762934B2 - Apparatus and method for verifying broadcast content object identification based on web data - Google Patents
Apparatus and method for verifying broadcast content object identification based on web data Download PDFInfo
- Publication number
- US9762934B2 US9762934B2 US14/932,426 US201514932426A US9762934B2 US 9762934 B2 US9762934 B2 US 9762934B2 US 201514932426 A US201514932426 A US 201514932426A US 9762934 B2 US9762934 B2 US 9762934B2
- Authority
- US
- United States
- Prior art keywords
- broadcast content
- web data
- knowledge
- content
- knowledge network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/951—Indexing; Web crawling techniques
-
- G06F17/30864—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/231—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
- H04N21/23109—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion by placing content in organized collections, e.g. EPG data repository
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234318—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into objects, e.g. MPEG-4 objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
- H04N21/2353—Processing of additional data, e.g. scrambling of additional data or processing content descriptors specifically adapted to content descriptors, e.g. coding, compressing or processing of metadata
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/237—Communication with additional data server
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/251—Learning process for intelligent management, e.g. learning user preferences for recommending movies
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/462—Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
- H04N21/4622—Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47205—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/4722—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting additional data associated with the content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/4728—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/482—End-user interface for program selection
- H04N21/4828—End-user interface for program selection for searching program descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/61—Network physical structure; Signal processing
- H04N21/6106—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
- H04N21/6125—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
- H04N21/8133—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
- H04N21/8405—Generation or processing of descriptive data, e.g. content descriptors represented by keywords
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8455—Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
Definitions
- the following description relates to broadcast communication technology, and particularly, to a technology for identifying an object in broadcast content.
- technologies used to object identification for broadcast content relate to image processing, whereby features are extracted from a specific scene (frame) of the broadcast content and an object with the features is selected from a group of candidate objects for identification.
- the current image processing technology only shows an average precision (AP) of 0.45, which is even decreased geometrically when the size of the group of objects to be identified increases.
- AP average precision
- the following description relates to an apparatus and method for verifying object identification, which are capable of verifying a result of object identification obtained by image processing, using information acquired from an external source, other than the broadcast content, and thereby increase the performance of object identification of broadcast content.
- an apparatus for verifying broadcast content object identification based on web data including: a web data processor configured to collect and process web data related to broadcast content and create content knowledge information by tagging the web data to the broadcast content; a content knowledge information storage portion configured to store the content knowledge information; and an object identification verifier configured to verify a result of identifying an object contained in the broadcast content, using the content knowledge information.
- a method for verifying broadcast content object identification based on web data including: collecting and processing web data related to broadcast content and creating content knowledge information by tagging the processed web data to the broadcast content; and verifying a result of identifying an object contained in the broadcast content, using the content knowledge information.
- FIG. 1 is a diagram illustrating an apparatus for verifying identification of a broadcast content object based on web data, according to an exemplary embodiment.
- FIG. 2 is a block diagram illustrating a web data processor according to an exemplary embodiment.
- FIG. 3 is a diagram illustrating an example of tagging a knowledge network to broadcast content.
- FIG. 4 is a diagram illustrating an object identification verifier according to an exemplary embodiment.
- FIG. 5 is a flowchart illustrating a method for verifying identification of a broadcast content object, based on web data, according to an exemplary embodiment.
- FIG. 6 is a flowchart illustrating the operation of generating content knowledge information by using web data according to the exemplary embodiment.
- FIG. 7 is a flowchart illustrating an operation of object identification verification according to the exemplary embodiment.
- the present disclosure suggests a technology that verifies object identification obtained by image processing, using information acquired from external sources, other than broadcast content.
- Broadcast content is delivered to a large number of audiences more promptly, compared to general videos or images, and the relevant content is frequently reproduced by the users. For example, in case of a South Korean television serial “My Love from the Star,” more than 40,000 relevant blog posts were generated during about a three-month period for which the show was aired.
- the present invention utilizes information extracted from the web data in order to verify object identification.
- FIG. 1 is a diagram illustrating an apparatus for verifying identification of a broadcast content object, based on web data, according to an exemplary embodiment.
- the apparatus 100 includes a web data processor 110 , a content knowledge information storage portion 120 , and an object identification verifier 130 .
- the web data processor 110 collects and processes web data related to broadcast content, and generates content knowledge information by tagging the broadcast content.
- the content knowledge information storage portion 120 stores content knowledge information, and the object identification verifier 130 verifies the identification of an object contained in the broadcast content, using the content knowledge information.
- a broadcast content database (DB) 10 includes a broadcast content-associated glossary 11 and broadcast content images 12 .
- the broadcast content DB 10 may be stored in a server of a broadcasting company, and it may be provided in various forms.
- the broadcast content DB 10 provides information related to broadcast content to the web data processor 110 , so that the web data processor 110 can create data to be used in verifying object identification by an object identifier 20 .
- the object identifier 20 identifies an object in a broadcast content image, using an image recognition technology.
- the object identification is verified by the object identification verifier 130 .
- FIG. 2 is a block diagram illustrating a web data processor according to an exemplary embodiment.
- the web data processor 110 includes, specifically, a web data collector 111 , a knowledge network constructor 112 , and a knowledge network tagger 113 .
- the web data collector 111 collects web data, using keywords relevant to an object. To this end, the web data collector 111 searches the broadcast content-associated glossary 11 to obtain keywords needed for collecting web data, creates a query by combining the obtained keywords, accesses a web portal 30 , and searches for and collects web data using the created query.
- the object-relevant keywords may include, for example, the title of content, a main character's name, the name of vehicle shown as product placement (PPL), or the name of location.
- the knowledge network constructor 112 constructs a knowledge network consisting of at least one of terms and images extracted from the collected web data.
- the knowledge network constructor 112 includes, specifically, an information/image extractor 112 a and a knowledge network creator 112 b .
- the information/image extractor 112 a extracts the terms and images from the web data (i.e., a web page).
- the knowledge network creator 112 b forms a knowledge network consisting of nodes that are the terms and images extracted by the information/image extractor 112 a .
- the knowledge network will be described in detail with reference to FIG. 3 .
- the knowledge network tagger 113 searches the broadcast content image to find a frame that contains an image which matches the image contained in the knowledge network, and tags the knowledge network to the found frame.
- the knowledge network tagger 113 includes, specifically, a matching model learner 113 a , a matching model creator 113 b , an image matcher 113 c , and a tagger 113 d.
- the matching model learner 113 a is trained to achieve a matching model on the basis of frames that form a broadcast content image 12 , which is stored in the broadcast content DB, and the matching model creator 113 b stores the trained matching model.
- the image matcher 113 c transforms feature vectors X into X′, wherein the feature vectors X are extracted from the broadcast content frame or the web images using the matching model, and measures similarity between the transformed vectors.
- the transformation of the feature vectors is carried out in such a manner that the similarity between broadcast content frames satisfies Equation 1 below as much as possible.
- Equation 1 sim( ) denotes a similarity calculation function, and #x denotes a frame number of x.
- the matching model that satisfies Equation 1 is learned in such a manner that the closer to each other the frames are, the more the similarity increases.
- the image matcher 113 c re-computes ranks of feature vectors to match, using a different feature extraction method from the previous feature extraction method used to build the feature vectors X. For example, if color information is used to extract feature vectors X, information about texture or boundaries of the images are used in the post-processing to build feature vectors Y, and the similarity of the feature vectors is computed, whereby n candidate objects obtained from the matching model have their ranks re-adjusted.
- Various linear, non-linear model learning methods such as SVMRank, manifold learning, and the like, may be applied to the establishment of matching model, and various feature extraction technologies may be used to extract features for the post-processing.
- the image matcher 113 c matches n frames with respect to one web image.
- the tagger 113 d may select one final frame of the highest score from among the n frames, tag the knowledge network to m frames preceding and following the final frame, and then assign weights to the m frames based on their distance to the final frame.
- the matching result and the knowledge network constructed by the knowledge network constructor 112 are combined with each other and the result is stored in the content knowledge information storage portion 120 .
- FIG. 3 is a diagram for explaining how to tag a knowledge network to broadcast content.
- the knowledge network G 310 consists of a plurality of nodes N 311 a and 311 b and edges E 312 that connect the nodes.
- each node N is a set of words 311 a and images 311 b that are extracted from collected web data.
- the edges E are formed based on the importance of information that corresponds to each node N in the web page and relevance between the information.
- the images 311 b contained in the knowledge network 310 may be matched with the frames 321 that form a broadcast content image 320 .
- FIG. 4 is a diagram illustrating an object identification verifier according to an exemplary embodiment.
- the object identification verifier 130 includes, specifically, a knowledge network extractor 131 , a graph search-based probability calculator 132 , a probability integrator 133 , an identification score re-adjuster 134 , and a final identification result outputter 135 .
- the knowledge network extractor 131 requests the content knowledge information storage portion 120 to send a set of knowledge networks C′ ⁇ C associated with the object-identified frame.
- the graph search-based probability calculator 132 calculates the appearance probability of an object based on a group O of candidate objects for identification and the set C′ of the knowledge networks. With respect to an object o ⁇ O to be identified, the appearance probability is calculated using Equation 2 as below.
- c ) ⁇ c ⁇ C ′ ⁇ ⁇ p ⁇ ( o
- #f denotes a frame number of the frame in which the object identification has been performed.
- the probability of an object o appearing in a knowledge network g c belonging to C′ is multiplied with a weight of the frame #f in each knowledge network. Such multiplication is performed for all c ⁇ C′.
- the probability integrator 133 computes the probability of an object o by summing the appearance probabilities calculated by the graph search-based probability calculator 132 .
- the identification score re-adjuster 134 re-adjusts the result of object identification from the object identifier 20 using the obtained probability of the object o, and performs identification verification.
- FIG. 5 is a flowchart illustrating a method for verifying identification of a broadcast content object, based on web data, according to an exemplary embodiment.
- the method for verifying identification of a broadcast content object based on web data mainly includes operations of: collecting and processing web data related to broadcast content and generating content knowledge information by tagging the broadcast content with the web data, as depicted in S 510 , and verifying identification of an object contained in the broadcast content by using the content knowledge information, as depicted in S 520 .
- FIG. 6 is a flowchart illustrating the operation of generating content knowledge information using web data according to the exemplary embodiment.
- the web data processor 110 collects the web data, as depicted in S 610 , constructs a knowledge network, as depicted in S 620 , and tags the knowledge network, as depicted in S 630 .
- web data is collected using object-relevant keywords for the broadcast content.
- the web data collector 111 searches the broadcast content-associated glossary 11 in an external broadcast content DB to acquire keywords needed for collecting web data, accesses a web portal 30 , and searches for and collects web data using the created query.
- the object-relevant keywords may include, for example, the title of content, a main character's name, the name of vehicle shown as product placement (PPL), or the name of location.
- a knowledge network consisting of at least one of images and terms extracted from the collected web data is constructed.
- operation S 620 includes extracting information/images, as depicted in S 621 , and constructing a knowledge network, as depicted in S 622 .
- terms and images are extracted from web data, i.e., web page.
- the knowledge network consisting of nodes that are the extracted terms and images is constructed.
- S 630 the broadcast content image is searched to find a frame that contains an image which matches the image contained in the knowledge network, and the knowledge network is tagged to the found frame.
- S 630 includes learning a matching model, as depicted in S 631 , storing the matching model, as depicted in S 632 , matching the image, as depicted in S 633 , and tagging the knowledge network to the frame, as depicted in S 634 .
- the matching model learner 113 a is trained to achieve a matching model based on frames that form the broadcast content image 12 stored in the broadcast content DB, as depicted in S 631 , and the matching model creator 113 b stores the trained matching model as depicted in S 632 .
- the image matcher 113 c matches n frames with respect to one web image, using the constructed knowledge network of S 622 and the stored matching model of S 632 . Specifically, the image matcher 113 c transforms feature vectors X into X′, wherein the feature vectors X are extracted from the broadcast content frame or the web images using the matching model, and measures similarity between the transformed vectors. At this time, the transformation of the feature vectors is carried out in such a manner that the similarity between broadcast content frames satisfies Equation 3 below as much as possible.
- x,y,z in X′ sim( x′,y′ ) ⁇ sim( x′,z′ ), if
- Equation 3 sim( ) denotes a similarity calculation function, and #x denotes a frame number of x.
- the matching model that satisfies Equation 3 is learned in such a manner that the closer to each other the frames are, the more the similarity increases.
- the image matcher re-computes ranks of feature vectors to match, using a different feature extraction method from the previous feature extraction method used to build the feature vectors X. For example, if color information is used to extract feature vectors X, information about texture or boundaries of the images are used in the post-processing to build feature vectors Y, and the similarity of the feature vectors is computed, whereby n candidate objects obtained from the matching model have their ranks re-adjusted.
- Various linear, non-linear model learning methods such as SVMRank, manifold learning, and the like, may be applied to the establishment of matching model, and various feature extraction technologies may be used to extract features for the post-processing.
- the tagger 113 d selects one final frame with the highest score from among the n frames, tags the knowledge network to m frames preceding and following the final frame, and then assigns weights to the m frames based on their distance to the final frame, as depicted in S 634 .
- FIG. 7 is a flowchart illustrating an operation of verifying object identification according to the exemplary embodiment.
- the operation of verifying object identification includes operations of: extracting a knowledge network as depicted in S 710 , computing probabilities based on graph search, as depicted in S 720 , integrating the probabilities as depicted in S 730 , re-adjusting the identification scores as depicted in S 740 , and outputting a final identification result as depicted in S 750 .
- the knowledge network extractor 131 requests the content knowledge information storage portion 120 to send a set of knowledge networks C′ ⁇ C associated with the object-identified frame, and receives the set, as depicted in S 710 .
- the graph search-based probability calculator 132 calculate the appearance probability of an object based on a group O of candidate objects for identification and the set C′ if knowledge networks. With respect to an object o ⁇ O to be identified, the appearance probability is calculated, as depicted in S 720 , using Equation 4 as below.
- c ) ⁇ c ⁇ C ′ ⁇ ⁇ p ⁇ ( o
- #f denotes a frame number of the frame in which the object identification has been performed.
- the probability of an object o appearing in a knowledge network g c belonging to C′ is multiplied with a weight of the frame #f in each knowledge network. Such multiplication is performed for all c ⁇ C′.
- the probability integrator 133 computes the probability of the object o by summing the appearance probabilities calculated by the graph search-based probability calculator 132 , as depicted in S 730 .
- the identification score re-adjuster 134 re-adjusts the result of object identification from the object identifier 20 using the obtained probability of the object o, and performs verification of the object identification, as depicted in S 740 , and outputs the final identification result, as depicted in S 750 .
- the apparatus and method according to the above exemplary embodiments can increase object recognition performance with respect to broadcast content, and hence be applicable to object-oriented search, recommendation or mashup.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- Computing Systems (AREA)
- Library & Information Science (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
sim(x′,y′)≦sim(x′,y′), if |(#x−#y)|≧|(#x−#z)|, for x,y,z in X′ (1)
for x,y,z in X′,sim(x′,y′)≦sim(x′,z′), if |(#x−#y)|≧|(#x−#z)| (3)
Claims (16)
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR20140152315 | 2014-11-04 | ||
| KR10-2014-0152315 | 2014-11-04 | ||
| KR1020150117556A KR101720685B1 (en) | 2014-11-04 | 2015-08-20 | Apparatus and Method for Web Data based Identification System for Object Identification Performance Enhancement |
| KR10-2015-0117556 | 2015-08-20 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20160127750A1 US20160127750A1 (en) | 2016-05-05 |
| US9762934B2 true US9762934B2 (en) | 2017-09-12 |
Family
ID=55854195
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/932,426 Expired - Fee Related US9762934B2 (en) | 2014-11-04 | 2015-11-04 | Apparatus and method for verifying broadcast content object identification based on web data |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US9762934B2 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11954150B2 (en) | 2018-04-20 | 2024-04-09 | Samsung Electronics Co., Ltd. | Electronic device and method for controlling the electronic device thereof |
Families Citing this family (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9762934B2 (en) * | 2014-11-04 | 2017-09-12 | Electronics And Telecommunications Research Institute | Apparatus and method for verifying broadcast content object identification based on web data |
| CN107832287A (en) * | 2017-09-26 | 2018-03-23 | 晶赞广告(上海)有限公司 | A kind of label identification method and device, storage medium, terminal |
| CN109960748B (en) * | 2019-02-22 | 2021-03-26 | 中国科学院计算技术研究所 | Method and system for prediction of individual live broadcast users watching anchors based on knowledge graph |
| KR20230030850A (en) * | 2021-08-26 | 2023-03-07 | 현대자동차주식회사 | Method and apparatus for providing broadcasting information based on machine learning |
Citations (18)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6507764B1 (en) * | 1999-03-30 | 2003-01-14 | Nct Group, Inc. | Network of digital broadcast stations |
| US20040103032A1 (en) * | 2000-10-12 | 2004-05-27 | Maggio Frank S. | Remote control system and method for interacting with broadcast content |
| US7047305B1 (en) * | 1999-12-09 | 2006-05-16 | Vidiator Enterprises Inc. | Personal broadcasting system for audio and video data using a wide area network |
| US20070107022A1 (en) * | 2004-06-08 | 2007-05-10 | Lawrence Bradley R Iii | Methods and apparatus to verify consumption of programming content |
| US20080229365A1 (en) * | 2007-03-12 | 2008-09-18 | Samsung Electronics Co., Ltd. | System, apparatus and method of providing personal broadcasting service according to scheduling |
| US20090217336A1 (en) * | 2008-02-22 | 2009-08-27 | Cyberlink Corp. | Playback Resume System and Method for a Media Center |
| KR100958302B1 (en) | 2008-06-19 | 2010-05-19 | 주식회사 아론티 | Object Information Provision Method in Data Broadcasting |
| US20120072420A1 (en) | 2010-09-16 | 2012-03-22 | Madhav Moganti | Content capture device and methods for automatically tagging content |
| US20130243307A1 (en) | 2012-03-16 | 2013-09-19 | Thomson Licensing | Object identification in images or image sequences |
| KR20140080724A (en) | 2012-12-14 | 2014-07-01 | 한국전자통신연구원 | Method for providing check-in service about object in multimedia |
| US20140211044A1 (en) | 2013-01-25 | 2014-07-31 | Electronics And Telecommunications Research Institute | Method and system for generating image knowledge contents based on crowdsourcing |
| KR20140095956A (en) | 2013-01-25 | 2014-08-04 | 한국전자통신연구원 | Method and system for generating image-knowledge contents based on crowdsourcing |
| US20140282673A1 (en) * | 2010-05-27 | 2014-09-18 | Zeev Neumeier | Systems and methods for real-time television ad detection using an automated content recognition database |
| US20140298375A1 (en) * | 2013-03-26 | 2014-10-02 | Kabushiki Kaisha Toshiba | Information output device, information manipulation device, and computer program product |
| US20140304597A1 (en) | 2013-04-05 | 2014-10-09 | Nbcuniversal Media, Llc | Content-object synchronization and authoring of dynamic metadata |
| US20160127750A1 (en) * | 2014-11-04 | 2016-05-05 | Electronics And Telecommunications Research Institute | Apparatus and method for verifying broadcast content object identification based on web data |
| US20160192029A1 (en) * | 2014-12-26 | 2016-06-30 | Mattias Bergstrom | Method and system for adaptive virtual broadcasting of digital content |
| US20160249115A1 (en) * | 2015-02-25 | 2016-08-25 | Yahoo! Inc. | Providing Interactivity Options for Television Broadcast Content |
-
2015
- 2015-11-04 US US14/932,426 patent/US9762934B2/en not_active Expired - Fee Related
Patent Citations (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6507764B1 (en) * | 1999-03-30 | 2003-01-14 | Nct Group, Inc. | Network of digital broadcast stations |
| US7047305B1 (en) * | 1999-12-09 | 2006-05-16 | Vidiator Enterprises Inc. | Personal broadcasting system for audio and video data using a wide area network |
| US20040103032A1 (en) * | 2000-10-12 | 2004-05-27 | Maggio Frank S. | Remote control system and method for interacting with broadcast content |
| US20070107022A1 (en) * | 2004-06-08 | 2007-05-10 | Lawrence Bradley R Iii | Methods and apparatus to verify consumption of programming content |
| US20080229365A1 (en) * | 2007-03-12 | 2008-09-18 | Samsung Electronics Co., Ltd. | System, apparatus and method of providing personal broadcasting service according to scheduling |
| US20090217336A1 (en) * | 2008-02-22 | 2009-08-27 | Cyberlink Corp. | Playback Resume System and Method for a Media Center |
| KR100958302B1 (en) | 2008-06-19 | 2010-05-19 | 주식회사 아론티 | Object Information Provision Method in Data Broadcasting |
| US20140282673A1 (en) * | 2010-05-27 | 2014-09-18 | Zeev Neumeier | Systems and methods for real-time television ad detection using an automated content recognition database |
| US20120072420A1 (en) | 2010-09-16 | 2012-03-22 | Madhav Moganti | Content capture device and methods for automatically tagging content |
| KR20130060299A (en) | 2010-09-16 | 2013-06-07 | 알까뗄 루슨트 | Content capture device and methods for automatically tagging content |
| US8533192B2 (en) | 2010-09-16 | 2013-09-10 | Alcatel Lucent | Content capture device and methods for automatically tagging content |
| US20130243307A1 (en) | 2012-03-16 | 2013-09-19 | Thomson Licensing | Object identification in images or image sequences |
| KR20140080724A (en) | 2012-12-14 | 2014-07-01 | 한국전자통신연구원 | Method for providing check-in service about object in multimedia |
| US20140211044A1 (en) | 2013-01-25 | 2014-07-31 | Electronics And Telecommunications Research Institute | Method and system for generating image knowledge contents based on crowdsourcing |
| KR20140095956A (en) | 2013-01-25 | 2014-08-04 | 한국전자통신연구원 | Method and system for generating image-knowledge contents based on crowdsourcing |
| US20140298375A1 (en) * | 2013-03-26 | 2014-10-02 | Kabushiki Kaisha Toshiba | Information output device, information manipulation device, and computer program product |
| US20140304597A1 (en) | 2013-04-05 | 2014-10-09 | Nbcuniversal Media, Llc | Content-object synchronization and authoring of dynamic metadata |
| US20160127750A1 (en) * | 2014-11-04 | 2016-05-05 | Electronics And Telecommunications Research Institute | Apparatus and method for verifying broadcast content object identification based on web data |
| US20160192029A1 (en) * | 2014-12-26 | 2016-06-30 | Mattias Bergstrom | Method and system for adaptive virtual broadcasting of digital content |
| US20160249115A1 (en) * | 2015-02-25 | 2016-08-25 | Yahoo! Inc. | Providing Interactivity Options for Television Broadcast Content |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11954150B2 (en) | 2018-04-20 | 2024-04-09 | Samsung Electronics Co., Ltd. | Electronic device and method for controlling the electronic device thereof |
Also Published As
| Publication number | Publication date |
|---|---|
| US20160127750A1 (en) | 2016-05-05 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9582518B2 (en) | Image processing apparatus, image processing method, and storage medium | |
| US9762934B2 (en) | Apparatus and method for verifying broadcast content object identification based on web data | |
| US20130148898A1 (en) | Clustering objects detected in video | |
| US9355330B2 (en) | In-video product annotation with web information mining | |
| JP5371083B2 (en) | Face identification feature value registration apparatus, face identification feature value registration method, face identification feature value registration program, and recording medium | |
| US10650274B2 (en) | Image clustering method, image clustering system, and image clustering server | |
| US20090259633A1 (en) | Universal Lookup of Video-Related Data | |
| US20130243307A1 (en) | Object identification in images or image sequences | |
| CN112380462A (en) | Method, device, server and computer readable storage medium for planning participation path | |
| CN107426610B (en) | Video information synchronization method and device | |
| US12354404B2 (en) | Obtaining artist imagery from video content using facial recognition | |
| CN113254665A (en) | Knowledge graph expansion method and device, electronic equipment and storage medium | |
| US9412049B2 (en) | Apparatus and method for recognizing object using correlation between object and content-related information | |
| CN118035489A (en) | Video searching method and device, storage medium and electronic equipment | |
| CN112101154A (en) | Video classification method and device, computer equipment and storage medium | |
| CN107247730A (en) | Image searching method and device | |
| CN118042217A (en) | Video processing method and device, electronic equipment and storage medium | |
| EP3096243A1 (en) | Methods, systems and apparatus for automatic video query expansion | |
| CN116451103B (en) | Situation element recommendation method based on label | |
| CN115408596A (en) | Object instance searching method, computing device and storage medium | |
| CN113408552A (en) | Feature quantization model training, feature quantization and data query methods and systems | |
| CN104850600A (en) | Method and device for searching images containing faces | |
| KR101720685B1 (en) | Apparatus and Method for Web Data based Identification System for Object Identification Performance Enhancement | |
| CN117217831B (en) | Advertisement putting method and device, storage medium and electronic equipment | |
| KR102228213B1 (en) | Method of recommendating image and apparatuses performing the same |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTIT Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SON, JEONG WOO;CHO, KEE SEONG;KIM, SUN JOONG;AND OTHERS;REEL/FRAME:036962/0138 Effective date: 20151028 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |
|
| FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
| FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20250912 |