[go: up one dir, main page]

CN106303726B - Video tag adding method and device - Google Patents

Video tag adding method and device Download PDF

Info

Publication number
CN106303726B
CN106303726B CN201610767921.XA CN201610767921A CN106303726B CN 106303726 B CN106303726 B CN 106303726B CN 201610767921 A CN201610767921 A CN 201610767921A CN 106303726 B CN106303726 B CN 106303726B
Authority
CN
China
Prior art keywords
label
client
video frame
tag
target video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610767921.XA
Other languages
Chinese (zh)
Other versions
CN106303726A (en
Inventor
邵明星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing QIYI Century Science and Technology Co Ltd
Original Assignee
Beijing QIYI Century Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing QIYI Century Science and Technology Co Ltd filed Critical Beijing QIYI Century Science and Technology Co Ltd
Priority to CN201610767921.XA priority Critical patent/CN106303726B/en
Publication of CN106303726A publication Critical patent/CN106303726A/en
Application granted granted Critical
Publication of CN106303726B publication Critical patent/CN106303726B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The embodiment of the invention discloses a method and a device for adding video tags, wherein the method comprises the following steps: receiving a tag adding instruction sent by a user through a first client in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and tag content of a tag adding target video frame; at least establishing an association relation among the target video, the video frame and the label content according to the identification information; and after receiving a request for acquiring the association relation sent by the second client, sending the association relation to the second client so that the second client displays the label content according to the association relation. By applying the embodiment of the invention, the user can add the video label.

Description

Video tag adding method and device
Technical Field
The invention relates to the technical field of video image processing, in particular to a method and a device for adding a video tag.
Background
With the development of video playing technology, playing videos stored in a server by a watching client through a network has become an indispensable part in the life of people nowadays. With the development of society, various communication demands among people on the internet are also expanding, and for watching videos, users generally like to forward and share the watched videos after watching the videos, and meanwhile, want to additionally share some ideas or other information about the videos. In this way, some good-looking videos or videos that the user likes can be shared with friends on the network.
When the network video is played, the label is displayed in the video to achieve the interaction with the user and the purpose of displaying the video information to the user, the mode is very effective, and the video information can be visually displayed to the user by combining the playing progress. The user can not set by himself, and can not show own idea or other information about the video to other users in a tag mode, and in order to provide more interaction modes or information display modes for the user, the user needs to add the tag by himself in the video playing process.
Disclosure of Invention
The embodiment of the invention aims to provide a method and a device for adding a video tag, so that a user can add the tag by himself through a client when watching a video played by the client.
In order to achieve the above purpose, the embodiment of the invention discloses a method and a device for adding a video tag. The technical scheme is as follows:
in a first aspect, a method for adding a video tag provided in an embodiment of the present invention is applied to a server, and the method includes:
receiving a tag adding instruction sent by a user through a first client in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and tag content of a tag adding target video frame;
at least establishing an association relation among the target video, the target video frame and the label content according to the identification information;
and after receiving a request sent by a second client for acquiring the association relationship, sending the association relationship to the second client so that the second client displays the label content according to the association relationship.
Preferably, the instruction further includes a tag display duration;
the establishing of the association relationship among the target video, the target video frame and the tag content at least according to the identification information includes:
and at least establishing the association relation among the target video, the target video frame, the label display duration and the label content according to the identification information.
Preferably, the instruction further comprises a label display position;
the establishing of the association relationship among the target video, the target video frame and the tag content at least according to the identification information includes:
and at least establishing the association relation among the target video, the target video frame, the label display position and the label content according to the identification information.
Preferably, the instruction further comprises a target image associated with the label;
at least establishing an association relation among the target video, the target video frame, the label display position and the label content according to the identification information; the method comprises the following steps:
establishing an association relation among the target video, the target video frame, the label display position and the label content according to the identification information;
determining video frames from the target video frame to a first video frame which is behind the target video frame and does not contain the target image by utilizing an image identification technology;
and establishing the incidence relation among the target video, the determined video frame, the label display position and the label content.
Preferably, the instruction further includes a target image, a target image position, and a tag display position;
at least establishing an association relation among the target video, the target video frame and the label content according to the identification information; the method comprises the following steps:
establishing an association relation among the target video, the target video frame, the label display position and the label content according to the identification information;
determining the corresponding relation between the target image position and the label display position according to the target image position and the label display position;
determining video frames from the target video frame to a first video frame which is behind the target video frame and does not contain the target image by utilizing an image identification technology;
determining the display position of the label in each video frame according to the corresponding relation aiming at each video frame in the determined video frames;
and establishing the incidence relation among the target video, the determined video frame, the determined display position and the label content.
Preferably, before sending the association relationship to the second client, the method further includes:
judging whether the incidence relation meets a preset requirement or not;
the sending the association relationship to the second client includes:
and sending the association relation to the second client side under the condition that the association relation meets a preset requirement.
Preferably, the method further comprises:
and deleting or forbidding the association relation under the condition that the association relation does not meet the preset requirement.
In a second aspect, an embodiment of the present invention provides an apparatus for adding a video tag, where the apparatus includes:
the instruction receiving module is used for receiving an instruction for adding a label, which is sent by a user through a first client, in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and label content of a target video frame to which the label is added;
the incidence relation establishing module is used for at least establishing the incidence relation among the target video, the target video frame and the label content according to the identification information;
and the sending module is used for sending the incidence relation to a second client after receiving a request for acquiring the incidence relation sent by the second client so that the second client can display the label content according to the incidence relation.
Preferably, the instruction further includes a tag display duration;
the association relationship establishing module is specifically configured to establish an association relationship among at least the target video, the target video frame, the tag display duration, and the tag content according to the identification information.
Preferably, the instruction further comprises a label display position;
the association relationship establishing module is specifically configured to establish an association relationship among at least the target video, the target video frame, the tag display position, and the tag content according to the identification information.
Preferably, the instruction further comprises a target image associated with the label;
the association relationship establishing module comprises:
a first association relationship establishing unit, configured to establish an association relationship among the target video, the target video frame, the tag display position, and the tag content according to the identification information;
a first determining unit, configured to determine, by using an image recognition technique, video frames between the target video frame and a first video frame that follows the target video frame and does not include the target image;
and the second incidence relation establishing unit is used for establishing incidence relations among the target video, the determined video frame, the label display position and the label content.
Preferably, the instruction further includes a target image, a target image position, and a tag display position;
the association relationship establishing module comprises:
a third association relationship establishing unit, configured to establish an association relationship among the target video, the target video frame, the tag display position, and the tag content according to the identification information;
the second determining unit is used for determining the corresponding relation between the target image position and the label display position according to the target image position and the label display position;
a third determining unit, configured to determine, by using an image recognition technique, video frames between the target video frame and a first video frame that follows the target video frame and does not include the target image;
a fourth determining unit, configured to determine, according to the correspondence, a display position of the tag in each of the determined video frames;
and the second incidence relation establishing unit is used for establishing incidence relations among the target video, the determined video frame, the determined display position and the label content.
Preferably, the apparatus further comprises:
the judging module is used for judging whether the incidence relation meets a preset requirement or not before sending the incidence relation to the second client;
the sending module is specifically configured to send the association relationship to the second client when the determination result of the determining module is yes.
Preferably, the apparatus further comprises:
and the filtering module is used for deleting or forbidding the association relation when the judgment result of the judging module is negative.
The embodiment of the invention discloses a method and a device for adding a video label, wherein a user sends a label adding instruction through a first client, the instruction at least comprises mark information and label content of a label adding target video frame, a server at least establishes an incidence relation among the target video, the video frame and the label content according to the content in the instruction after receiving the instructions, and sends the incidence relation to a second client after receiving a request for acquiring the incidence relation sent by the second client, so that the second client displays the corresponding label content according to the incidence relation, and the function of automatically adding the label through the client when the user watches the video through the client can be realized. Therefore, the designability of the video label is increased, and the diversity of user interaction modes and the diversity of information display during video playing are increased.
Of course, it is not necessary for any product or method of practicing the invention to achieve all of the above-described advantages at the same time.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a schematic flowchart of a method for adding a video tag according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of an apparatus for adding a video tag according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
In order to enable a user to add a label to a network video played by the user through a client, the embodiment of the invention provides a method and a device for adding a video label.
It should be noted that the client applied in the embodiment of the present invention refers to different network video playing software installed on a terminal device, where the terminal device may be a mobile terminal, the mobile terminal may be a notebook, a tablet computer, a mobile phone, and the like, and the terminal device may also be a fixed terminal, such as an intelligent television, a desktop computer, and the like, which are all reasonable.
First, a method for adding a video tag provided by an embodiment of the present invention is described below.
As shown in fig. 1, fig. 1 is a schematic flowchart of a method for adding a video tag according to an embodiment of the present invention, where the method for adding a video tag is applied to a server, and specifically includes the following steps:
s101, receiving a label adding instruction sent by a user through a first client in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and label content of a target video frame to which a label is added;
in specific implementation, the entry for sending the instruction for adding the tag may be a tag adding button set at the first client, and the user triggers a behavior of adding the tag by himself by clicking the button, and can enter an editing interface of the video tag after clicking the button, and at least can edit the identification information of the target video frame to which the tag is added and the tag content, and of course, the entry for sending the instruction for adding the tag is set, and the triggering mode is also set by other modes.
The target video frame to which the tag is added may be a current frame when the user sends a tag adding instruction through the first client, or may be a certain frame specified by the user through setting; the identification information is used to mark the corresponding video frame so that the corresponding video frame can be identified by the identification information, and specifically, the identification information may be a video frame time point, a video frame ID, and the like, but is not limited thereto.
It should be noted that, at any time when the video is played, the user may send an instruction to add a tag through the first client, so that the user can add the tag in the target video by himself.
In specific implementation, the instruction may further include a tag display duration; or a label display location; or a tag display position, a target image associated with the tag; or target image, target image position, and tag display position, further precise control of the added tag may be achieved.
S102, at least establishing an association relation among the target video, the target video frame and the label content according to the identification information;
in order to accurately display the tag content, the target video frame to which the tag is added, and the tag content need to be accurately associated with each other, and therefore, an association relationship between them needs to be established.
When the step is implemented specifically, the corresponding association relationship is established according to the content included in the instruction, which is specifically as follows:
if the instruction comprises the identification information of the target video frame added with the label and the label content, at least the association relation of the target video, the target video frame and the label content can be established according to the identification information.
Illustratively, when a user watches a runner in the second period of the fourth season, when a video is played to the 1 st S, a command for adding a label is sent to a server through a first client, wherein the command comprises the 1 st S of mark information of a target video frame and label content 'deer break garment good', and an established association relation 1 is established; when the video is played to the 10 th S, sending a label adding instruction to the server through the first client, wherein the instruction comprises the 10 th S of the mark information of the target video frame and the label content 'the clothes with the break down can be purchased from the Teeta', and establishing an association relation 2; when the video is played to the 20 th S, sending a tag adding instruction to the server through the first client, wherein the instruction comprises the 20 th S of the mark information of the target video frame and the tag content of 'clothes in the morning are not good at sight', and establishing an association relation 3. Similarly, more associations may be established. The association established in this example is shown in table 1:
TABLE 1
Target video Video frame Label content
Association relation 1 The second stage of the fourth season for male race 1 st S "deer break clothes good"
Association 2 The second stage of the fourth season for male race 10S "deer break down clothes can be purchased from tianmao"
Association relation 3 The second stage of the fourth season for male race 20 th S ' Li Chen ' clothes are not good-looking '
If the instruction comprises identification information of the target video frame added with the label, label content and label display duration, at least the association relation between the target video, the target video frame, the label display duration and the label content can be established according to the identification information.
Illustratively, when a user watches a runner in the second period of the fourth season, when a video is played to the 1 st S, a command for adding a label is sent to a server through a first client, the 1 st S of mark information of a target video frame, the label content 'deer break and clothing error' and the label display duration 3S in the command form an established association relationship 1; when the video is played to the 10 th S, sending a label adding instruction to the server through the first client, wherein the instruction comprises the 10 th S of the mark information of the target video frame, the label content ' the clothes can be purchased from the cat ' and the label display time length is 5S ', and establishing an association relation 2; when the video is played to the 20 th S, a label adding instruction is sent to the server through the first client, the instruction comprises the 20 th S of the mark information of the target video frame, the label content of 'clothes in the morning are not good at sight' and the label display time length 11S, and the association relation 3 is established. Similarly, more associations may be established. The association established in this example is shown in table 2:
TABLE 2
Figure BDA0001099333120000081
If the instruction comprises the identification information of the target video frame added with the label, the label content and the label display position, at least the association relation among the target video, the target video frame, the label display position and the label content can be established according to the identification information.
It should be noted that the label display position may be rectangular, or may be in other shapes, and specifically may be determined by recording the coordinates of multiple points on the edge of the position. For example, in the case of a rectangle, the display position of the rectangular tag can be determined by coordinates of four corner points of the rectangle.
Illustratively, when a user watches a runner in the second period of the fourth season, when the video is played to the 1 st S, a command for adding a label is sent to the server through the first client, the command includes the 1 st S of mark information of a target video frame, a label content of 'deer break clothes good' and a label display position, the label display position is a rectangular area determined by coordinate points (0, 0), (0, 10), (3, 0) and (3, 10) which are directly input, or a rectangular area selected by a mouse on a playing interface of the first client, the first client recognizes that coordinates of four corner points are (0, 0), (0, 10), (3, 0) and (3, 10), and the established association relationship is 1; when the video is played to the 10 th S, sending a label adding instruction to the server through the first client, wherein the instruction comprises the 10 th S of the mark information of the target video frame, the label content 'the clothes with the deer break can be purchased from a cat' and a label display position, the label display position is a rectangular area determined by coordinate points (0, 10), (0, 20), (3, 10), (3, 20) which are directly input, or a rectangular area selected by a mouse on a playing interface of the first client, the first client identifies that the coordinates of four corner points are respectively (0, 10), (0, 20), (3, 10), (3, 20), and establishes an association relation 2; when the video is played to the 20 th S, sending a tag adding instruction to the server through the first client, wherein the instruction comprises the 20 th S of the mark information of the target video frame, the tag content 'poor clothes in the morning' and the tag display position, the tag display position is a rectangular area determined by coordinate points (10, 10), (10, 20), (13, 10), (13, 20) which are directly input, or a rectangular area selected by a mouse on a playing interface of the first client, the first client identifies that the coordinates of four corner points are (10, 10), (10, 20), (13, 10), (13, 20) respectively, and an association relation 3 is established; similarly, more associations may be established. The association established in this example is shown in table 3:
TABLE 3
Figure BDA0001099333120000091
If the instruction comprises identification information of a target video frame added with a label, label content, a label display position and a target image associated with the label, at least establishing an association relation among the target video, the target video frame, the label display position and the label content according to the identification information, and the method comprises the following steps:
(1) according to the identification information, establishing an incidence relation among the target video, the target video frame, the label display position and the label content;
(2) determining a video frame from a target video frame to a first video frame which is behind the target video frame and does not contain the target image by utilizing an image identification technology;
(3) and establishing an incidence relation among the target video, the determined video frame, the label display position and the label content.
It should be noted that the label display position may be rectangular, or may be in other shapes, and specifically may be determined by recording the coordinates of multiple points on the edge of the position. For example, in the case of a rectangle, the display position of the rectangular tag can be determined by coordinates of four corner points of the rectangle.
Illustratively, when a user watches a runner in the second fourth season, when the video is played to the 1 st S, a command for adding a label is sent to the server through the first client, the command includes the 1 st S of the mark information of the target video frame, a label content "a cap broken by deer can be purchased in kyoto", a label display position, and a "cap broken by deer" target image related to the label, the label display position is a rectangular region determined by coordinate points (0, 0), (0, 10), (3, 0), (3, 10) which are directly input, or a rectangular region selected by a mouse on a first client playing interface frame, the first client recognizes that coordinates of four corner points are (0, 0), (0, 10), (3, 0), (3, 10), and establishes an association relationship 1; the server determines video frames containing a target image of 'a hat broken down' by using an image recognition technology, wherein the video frames comprise a first frame, a second frame and a third frame after the 1 st S, and a fourth frame does not contain the target image, and an association relation 2, an association relation 3 and an association relation 4 are respectively established for the three frames. Similarly, more associations may be established. The association established in this example is shown in table 4:
TABLE 4
Figure BDA0001099333120000101
If the instruction includes identification information of the target video frame added with the label, label content, a target image position and a label display position, establishing the association relationship comprises:
(1) according to the identification information, establishing an incidence relation among the target video, the target video frame, the label display position and the label content;
(2) determining the corresponding relation between the target image position and the label display position according to the target image position and the label display position;
(3) determining a video frame from a target video frame to a first video frame which is behind the target video frame and does not contain the target image by utilizing an image identification technology;
(4) determining the display position of the label in the video frame according to the corresponding relation aiming at each video frame in the determined video frames;
(5) and establishing the association relation among the target video, the determined video frame, the determined display position and the label content.
Illustratively, when the user watches the runner in the second year of the fourth season, when the video is played to the 1 st S, an instruction for adding a label is sent to the server through the first client, the instruction includes the identification information 1 st S of the target video frame with the label, the label content "the hat is broken can be purchased in the kyoto", the target image "the hat is broken", the target image position, and the label display position, the target image position may be a position of any point on the hat, such as the center position of the hat, the coordinates are (10, 2), the label display position is a rectangular area determined by four coordinate points (12, 2), (22, 2), (12, 5), (22, 5) on the right side of the center position, the first video frame after the target video frame and not containing the target image is determined to be the third frame after the 1 st S by using an image recognition technology, and therefore, determining a first frame and a second frame after the 1S, wherein the central position of the target image in the first frame is (10, 1), and determining the display positions of the labels in the first frame in the video frame to be (12, 1), (22, 1), (12, 4), (22, 4) according to the corresponding relation between the central position of the original hat and the display positions of the labels; if the central position of the target image in the second frame is (20, 5), the display positions of the corresponding determination labels in the video frame are (22, 5), (32, 5), (22, 8) and (32, 8), and an association relation 2 and an association relation 3 are respectively established for the two frames; similarly, more associations may be established. The association established in this example is shown in table 5:
TABLE 5
Figure BDA0001099333120000111
Figure BDA0001099333120000121
S103, after receiving a request for acquiring the association relation sent by the second client, sending the association relation to the second client so that the second client can display the label content according to the association relation;
the second client may send all the association relations of the acquired target video when the target video starts to be played, or may send the association relations of the acquired target video in the next preset time period at intervals of a preset time period.
In the specific implementation of this step, after the second client receives the association relationship, when the second client plays the target video and reaches the video frame corresponding to any association relationship, the tag content corresponding to any association relationship is displayed according to any association relationship. The second client displays the label content, which specifically includes the following steps:
if the association relationship among the target video, the target video frame, and the tag content is established in step S102, after the second client obtains the association relationship, when the second client plays the target video to the corresponding target video frame, the tag content is displayed, specifically at what position, and the second client may randomly designate or may be preset.
For example, the association relationship shown in table 1 is established, after the second client obtains the association relationship in table 1, when the second client plays from the second period to the first period from the fourth season for men, according to the association relationship 1, when the second client plays from the target video to the target video frame corresponding to the first period, the second client displays that the label content is "deer break clothing is good"; when the client plays from the second period to the 10 th S of the fourth season of a male, according to the association relation 2, when the second client plays from the target video to the target video frame corresponding to the 10 th S, the second client displays that the label content is ' deer break ' clothes can be purchased from a sky cat '; when the client plays from the second stage to the 20 th S of the fourth season of the male, the second client displays the label content as 'unsightly clothing in the morning' according to the association relation 3 when the second client plays the target video to the target video frame corresponding to the 20 th S.
If the association relationship among the target video, the target video frame, the tag display duration and the tag content is established in step S102, after the second client obtains the association relationship, when the second client plays the target video to the corresponding target video frame, the second client displays the tag content according to the association relationship, the displayed duration is the tag display duration, and after the tag display duration, the tag content is hidden, so that the user can also set the time length for tag display by himself.
For example, the association relationship shown in table 2 is established, after the second client obtains the association relationship in table 2, when the second client plays from the second period of the fourth season to the 1 st S of a runner, according to the association relationship 1, the second client displays the label content "deer break and clothing is not wrong", and displays 3S, and after 3S, the label is automatically hidden; when the second client plays from the second season to the 10 th season of a runner, according to the association relation 2, the second client displays the label content that ' the clothes can be purchased from a skatecat if the deer break ' occurs ', and displays 5S, and the label is automatically hidden after 5S; when the second client plays from the second season to the 20 th season, the second client displays the label content 'the clothes in the morning are not good at sight' according to the association relation 3, 11S is displayed, and the label is automatically hidden after 11S.
If the association relationship among the target video, the target video frame, the tag display position and the tag content is established in step S102, the second client displays the tag content at the tag display position according to the association relationship, so that the user can add the tag at the position where the user wants to add the tag.
For example, the association relationship shown in table 3 is established, after the second client obtains the association relationship in table 3, when the client plays from the second season to the 1 st S of the fourth season for a male, according to the association relationship 1, the second client displays the label content "deer break and clothing good" within the rectangular range determined by the coordinate points (0, 0), (0, 10), (3, 0) and (3, 10); when the client plays from the second stage to the 10 th S of the fourth season of the male, the second client displays the label content 'deer break and clothing is good' in the rectangular range determined by the coordinate points (0, 10), (0, 20), (3, 10) and (3, 20) according to the association relation 2; when the client plays from the second to the 20 th season of the male, the second client displays the label content 'clothes in plum morning' in the rectangular range determined by the coordinate points (10, 10), (10, 20), (13, 10), (13, 20) according to the association relation 3.
If the association relationship among the target video, the target video frame, the tag display position and the tag content and the association relationship among the target video, the determined video frames in which the target images continuously appear, the tag display position and the tag content are established in step S102, the second client displays the tag content on the tag display position of the video frame corresponding to the association relationship.
For example, the association relationship shown in table 4 above is established, and after the second client obtains the association relationship in table 4, when the second client plays the second period from the fourth to the 1 st S, the first frame after the 1 st S, the second frame after the 1 st S, and the third frame after the 1 st S, the second client displays the label content "deer down hat can be purchased in the kyotong" within the rectangular range determined by the coordinate points (0, 0), (0, 10), (3, 0), (3, 10) of the frames according to the association relationship 1, the association relationship 2, the association relationship 3, and the association relationship 4, respectively.
If the step S102 establishes the incidence relation among the target video, the video frame, the label display position and the label content; and the second client can display the label content at the display position of the label in the video frame corresponding to the association relationship according to the association relationship, so that the addition of a dynamic label can be realized, and the label content can move along with the movement of the target image.
For example, the association relationship shown in table 5 is established, and after the second client obtains the association relationship in table 5, when the second client plays from the second season to the 1 st S of the fourth season of the runner, the second client displays the label content "the hat broken by deer can be purchased in the kyoto" within the rectangular range determined by the coordinate points (12, 2), (22, 2), (12, 5), (22, 5) according to the association relationship 1; when the child hat is played to the first frame after the 1 st S, according to the association relation 2, the label content "the child hat can be purchased in the Jingdong" is displayed within the rectangular range determined by the coordinate points (12, 1), (22, 1), (12, 4), (22, 4); when playing is carried out to the first frame after the 3 rd frame, according to the association relation 2, the label content "the cap that deer break can be purchased in the east of Beijing" is displayed in the rectangular range determined by the coordinate points (22, 5), (32, 5), (22, 8) and (32, 8), thereby realizing the addition of the dynamic label.
By applying the embodiment shown in fig. 1, a user adds a tag instruction through a client, the instruction at least includes tag information and tag content of a target video frame to which the tag is added, the server at least establishes an association relationship between the target video, the video frame and the tag content according to the content in the instruction after receiving the instruction, and sends the association relationship to a second client after receiving a request for acquiring the association relationship sent by the second client, so that the second client displays the corresponding tag content according to the association relationship, thereby realizing a function of automatically adding the tag through the client when the user watches the video through the client. Therefore, the designability of the label is increased, and the diversity of user interaction modes and the diversity of information display during video playing are increased.
In addition, in order to enhance the diversity of video playing and user interaction, it may also be arranged to trigger a jump to a link related to the tag content by clicking the tag content, for example, if the tag content is related to a pair of glasses displayed in the video, the link related to the tag content may be a website for purchasing the glasses, or a related knowledge introduction of the glasses, etc.
In this embodiment, before sending the association relationship to the second client, the method provided in this embodiment may further include:
firstly, judging whether the incidence relation meets the preset requirement or not;
in specific implementation, if the association relationship established for a certain frame is too many, when the second client plays the video to the frame, if the content of the tag is displayed according to each association relationship of the frame, too many tags are displayed in the video frame, and even the video frame is full of the tag, which gives a user a bad impression. In this case, some of the associations need to be filtered out, and the associations are filtered by setting conditions. Some user-added tags may be filtered out based on the type of user, for example, if a VIP user adds, the association is deemed to satisfy a preset requirement, and if not, the association is deemed to be not satisfied. Of course, the number of the added additives can be limited according to the number, the number of the added additives in the former preset number meets the preset requirement, and the number of the added additives in the later preset number does not meet the preset requirement.
Then, sending the association relation to a second client, including:
and sending the association relation to the second client side under the condition that the association relation meets the preset requirement.
Specifically, for the association relation which does not meet the preset requirement, a deletion or a disabling process may be adopted.
Under the condition, only the incidence relation meeting the preset requirement is sent to the second client side, and the second client side displays the label content according to the received incidence relation, so that the problem that the label content is displayed too much can be solved.
Corresponding to the above method embodiment, an embodiment of the present invention further provides an apparatus for adding a video tag, as shown in fig. 2, fig. 2 is a schematic structural diagram of the apparatus for adding a video tag according to the embodiment of the present invention, and the apparatus may include the following modules:
the instruction receiving module 201 is configured to receive an instruction for adding a tag, which is sent by a user through a first client, in a process of playing a target video by the first client, where the instruction at least includes identification information of a target video frame to which the tag is added and tag content;
an association relationship establishing module 202, configured to at least establish an association relationship between the target video, the target video frame, and the tag content according to the identification information;
the sending module 203 is configured to send the association relationship to the second client after receiving the request for obtaining the association relationship sent by the second client, so that the second client displays the tag content according to the association relationship.
By applying the embodiment shown in fig. 2, a user sends a tag adding instruction through a client, the instruction at least includes tag information and tag content of a target video frame to which a tag is added, after receiving the instructions, a server at least establishes an association relationship between the target video, the video frame and the tag content according to the content in the instruction, and after receiving a request for acquiring the association relationship sent by a second client, sends the association relationship to the second client, so that the second client displays the corresponding tag content according to the association relationship, thereby realizing a function of automatically adding a tag through the client when the user watches a video through the client, increasing designability of a video tag, and increasing diversity of user interaction modes and diversity of information display when the video is played.
In this embodiment, the association relationship establishing module 202 establishes a corresponding association relationship according to the content included in the instruction received by the instruction receiving module 201, specifically as follows:
if the instruction at least comprises the identification information of the target video frame added with the label, the label content and the label display duration;
the association relationship establishing module 202 is specifically configured to establish an association relationship among at least a target video, a target video frame, a tag display duration, and a tag content according to the identification information.
If the instruction at least comprises the identification information of the target video frame added with the label, the label content and the label display position;
the association relationship establishing module 202 is specifically configured to establish an association relationship among at least a target video, a target video frame, a tag display position, and tag content according to the identification information.
If the instruction at least comprises identification information of the video frame to which the label is added, the content of the label, the display position of the label and a target image associated with the label;
the association relationship establishing module 202 includes:
a first association relationship establishing unit (not shown in the figure) for establishing an association relationship among the target video, the target video frame, the tag display position and the tag content according to the identification information;
a first determining unit (not shown in the figure) for determining a video frame from the target video frame to a first video frame which is subsequent to the target video frame and does not contain the target image by using an image recognition technology;
and a second association relationship establishing unit (not shown in the figure) for establishing an association relationship among the target video, the determined video frame, the tag display position and the tag content.
If the instruction at least comprises the identification information of the video frame added with the label, the label content, the target image, the position of the target image and the display position of the label;
the association relationship establishing module 202 includes:
a third association relationship establishing unit (not shown in the figure) for establishing an association relationship among the target video, the target video frame, the tag display position and the tag content according to the identification information;
a second determining unit (not shown in the figure) for determining a corresponding relationship between the target image position and the label display position according to the target image position and the label display position;
a third determining unit (not shown in the figure) for determining a video frame from the target video frame to a first video frame which is subsequent to the target video frame and does not contain the target image by using an image recognition technology;
a fourth determining unit (not shown in the figure) for determining, for each of the determined video frames, a display position of the tag in the video frame according to the corresponding relationship;
and a second association relationship establishing unit (not shown in the figure) for establishing an association relationship among the target video, the determined video frame, the determined display position and the tag content.
It should be noted that the second client displays the tag content according to the obtained specific content of the association relationship. The method comprises the following specific steps:
if the association relationship between the target video, the video frame and the tag content is established by the association relationship establishing module 202, the second client displays the tag content after the sending module 203 sends the association relationship to the second client.
If the association relationship between the target video, the video frame, the tag display duration and the tag content is established by the association relationship establishing module 202, after the sending module 203 sends the association relationship to the second client, the second client displays the video tag content, the display duration is the tag display duration, and after the tag display duration, the tag content can be hidden and disappear, so that the user can also set the time length for tag display by himself.
If the association relationship among the target video, the target video frame, the tag display position and the tag content is established by the association relationship establishing module 202, after the sending module 203 sends the association relationship to the second client, the second client displays the tag at the tag display position, so that the user can add the tag at the position where the user wants to add the tag, specifically, the tag display position can be obtained by first selecting the desired position in the target video frame by the user and then calculating by the first client, and the tag display position is input into the tag adding instruction.
If the association relationship between the target video, the target video frame, the tag display position, and the tag content and the association relationship between the target video, the video frame where the determined target image appears, the tag display position, and the tag content are established by the association relationship establishing module 202, after the sending module 203 sends the association relationship to the second client, the second client will display the tag content at the tag display position of the target video frame and the video frame where the determined target image appears, so that the user can add the tag in the frame where the target image appears.
If the association relationship establishing module 202 establishes the association relationship among the target video, the video frame, the label display position and the label content; and the association relationship among the target video, the determined video frame in which the target image appears, the determined display position, and the tag content, after the sending module 203 sends the association relationship to the second client, the second client may display the tag content at the display position of the corresponding tag in the target video frame and the determined video frame in which the target image appears.
In this embodiment, the method may further include: the judging module is used for judging whether the incidence relation meets the preset requirement or not before sending the incidence relation to the second client;
the sending module 203 is specifically configured to send the corresponding association relationship to the second client when the determination result of the determining module is yes.
Further, this embodiment may further include:
and a filtering module (not shown in the figure) for deleting or disabling the association relationship when the association relationship does not meet the preset requirement.
Further, this embodiment may further include:
and a link module (not shown in the figure) for connecting the client to the link after receiving the instruction for triggering the jump to the link related to the label content. Specifically, the user may trigger the instruction by sliding or clicking the content of the tag displayed by the second client.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
All the embodiments in the present specification are described in a related manner, and the same and similar parts among the embodiments may be referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, as for the apparatus embodiment, since it is substantially similar to the method embodiment, the description is relatively simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
Those skilled in the art will appreciate that all or part of the steps in the above method embodiments may be implemented by a program to instruct relevant hardware to perform the steps, and the program may be stored in a computer-readable storage medium, which is referred to herein as a storage medium, such as: ROM/RAM, magnetic disk, optical disk, etc.
The above description is only for the preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention shall fall within the protection scope of the present invention.

Claims (6)

1. A method for adding video tags is applied to a server, and comprises the following steps:
receiving a tag adding instruction sent by a user through a first client in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and tag content of a tag adding target video frame;
at least establishing an association relation among the target video, the target video frame and the label content according to the identification information;
after receiving a request sent by a second client for acquiring the incidence relation, sending the incidence relation to the second client so that the second client displays the label content according to the incidence relation;
the instruction further comprises a target image, a target image position and a label display position, wherein the label display position is an area for displaying the label content, and the target image position is the position of any point in the target image;
at least establishing an association relation among the target video, the target video frame and the label content according to the identification information; the method comprises the following steps:
establishing an association relation among the target video, the target video frame, the label display position and the label content according to the identification information;
determining a relative position relation between the target image position and the label display position according to the target image position and the label display position;
determining video frames from the target video frame to a first video frame which is behind the target video frame and does not contain the target image by utilizing an image identification technology;
for each video frame in the determined video frames, determining the display position of the label in the video frame according to the relative position relationship, wherein the relative position relationship is formed between the determined display position and the position of the target image in the video frame;
and establishing the incidence relation among the target video, the determined video frame, the determined display position and the label content.
2. The method of claim 1, wherein before sending the association to the second client, the method further comprises:
judging whether the incidence relation meets a preset requirement or not;
the sending the association relationship to the second client includes:
and sending the association relation to the second client side under the condition that the association relation meets a preset requirement.
3. The method of claim 2, further comprising:
and deleting or forbidding the association relation under the condition that the association relation does not meet the preset requirement.
4. An apparatus for adding a video tag, the apparatus comprising:
the instruction receiving module is used for receiving an instruction for adding a label, which is sent by a user through a first client, in the process of playing a target video by the first client, wherein the instruction at least comprises identification information and label content of a target video frame to which the label is added;
the incidence relation establishing module is used for at least establishing the incidence relation among the target video, the target video frame and the label content according to the identification information;
the sending module is used for sending the incidence relation to a second client after receiving a request for acquiring the incidence relation sent by the second client so that the second client can display the label content according to the incidence relation;
the instruction further comprises a target image, a target image position and a label display position, wherein the label display position is an area for displaying the label content, and the target image position is the position of any point in the target image;
the association relationship establishing module comprises:
a third association relationship establishing unit, configured to establish an association relationship among the target video, the target video frame, the tag display position, and the tag content according to the identification information;
the second determining unit is used for determining the relative position relationship between the target image position and the label display position according to the target image position and the label display position;
a third determining unit, configured to determine, by using an image recognition technique, video frames between the target video frame and a first video frame that follows the target video frame and does not include the target image;
a fourth determining unit, configured to determine, for each of the determined video frames, a display position of the tag in the video frame according to the relative position relationship, where the determined display position and a position of a target image in the video frame have the relative position relationship;
and the second incidence relation establishing unit is used for establishing incidence relations among the target video, the determined video frame, the determined display position and the label content.
5. The apparatus of claim 4, further comprising:
the judging module is used for judging whether the incidence relation meets a preset requirement or not before sending the incidence relation to the second client;
the sending module is specifically configured to send the association relationship to the second client when the determination result of the determining module is yes.
6. The apparatus of claim 5, further comprising:
and the filtering module is used for deleting or forbidding the association relation when the judgment result of the judging module is negative.
CN201610767921.XA 2016-08-30 2016-08-30 Video tag adding method and device Active CN106303726B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610767921.XA CN106303726B (en) 2016-08-30 2016-08-30 Video tag adding method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610767921.XA CN106303726B (en) 2016-08-30 2016-08-30 Video tag adding method and device

Publications (2)

Publication Number Publication Date
CN106303726A CN106303726A (en) 2017-01-04
CN106303726B true CN106303726B (en) 2021-04-16

Family

ID=57675224

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610767921.XA Active CN106303726B (en) 2016-08-30 2016-08-30 Video tag adding method and device

Country Status (1)

Country Link
CN (1) CN106303726B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104602128A (en) * 2014-12-31 2015-05-06 北京百度网讯科技有限公司 Video processing method and device
CN108628913B (en) * 2017-03-24 2024-06-25 上海交通大学 Video processing method and device
CN109274926B (en) * 2017-07-18 2020-10-27 杭州海康威视系统技术有限公司 Image processing method, device and system
CN108271064B (en) * 2018-01-08 2021-03-16 武汉斗鱼网络科技有限公司 An anchor search method, device, electronic device and medium
CN109040823B (en) * 2018-08-20 2021-06-04 青岛海信传媒网络技术有限公司 Bookmark display method and device
CN110874167B (en) * 2018-08-29 2023-04-21 阿里巴巴集团控股有限公司 Data processing method, apparatus and machine readable medium
CN109391849B (en) * 2018-09-30 2020-11-20 联想(北京)有限公司 Processing method and system, multimedia output device and memory
CN109600635B (en) * 2019-01-24 2021-04-27 四川长虹电器股份有限公司 System and method for pushing information according to intelligent television video on demand
CN110035249A (en) * 2019-03-08 2019-07-19 视联动力信息技术股份有限公司 A kind of video gets method and apparatus ready
CN110730382B (en) * 2019-09-27 2020-10-30 北京达佳互联信息技术有限公司 Video interaction method, device, terminal and storage medium
CN111491206B (en) * 2020-04-17 2023-03-24 维沃移动通信有限公司 Video processing method, video processing device and electronic equipment
CN115361596A (en) * 2022-07-04 2022-11-18 浙江大华技术股份有限公司 Panoramic video data processing method, device, electronic device and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104244111A (en) * 2013-06-20 2014-12-24 深圳市快播科技有限公司 Method and device for labeling medium attribute of video
CN104317490A (en) * 2014-09-30 2015-01-28 联想(北京)有限公司 Promoting method and electronic equipment
CN105052155A (en) * 2013-03-20 2015-11-11 谷歌公司 Interpolated video tagging
CN105872820A (en) * 2015-12-03 2016-08-17 乐视云计算有限公司 Method and device for adding video tag

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8761574B2 (en) * 2012-10-04 2014-06-24 Sony Corporation Method and system for assisting language learning
CN103780973B (en) * 2012-10-17 2017-08-04 三星电子(中国)研发中心 Video tab adding method and device
US9407975B2 (en) * 2013-03-05 2016-08-02 Brandon Grusd Systems and methods for providing user interactions with media
CN103970906B (en) * 2014-05-27 2017-07-04 百度在线网络技术(北京)有限公司 The method for building up and device of video tab, the display methods of video content and device
CN104967908B (en) * 2014-09-05 2018-07-24 腾讯科技(深圳)有限公司 Video hotspot labeling method and device
CN104391960B (en) * 2014-11-28 2019-01-25 北京奇艺世纪科技有限公司 A video annotation method and system
CN105072460B (en) * 2015-07-15 2018-08-07 中国科学技术大学先进技术研究院 A kind of information labeling and correlating method based on video content element, system and equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105052155A (en) * 2013-03-20 2015-11-11 谷歌公司 Interpolated video tagging
CN104244111A (en) * 2013-06-20 2014-12-24 深圳市快播科技有限公司 Method and device for labeling medium attribute of video
CN104317490A (en) * 2014-09-30 2015-01-28 联想(北京)有限公司 Promoting method and electronic equipment
CN105872820A (en) * 2015-12-03 2016-08-17 乐视云计算有限公司 Method and device for adding video tag

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"细粒度视频标签机制及其应用研究";曲鑫;《中国优秀硕士学位论文全文数据库》;20130715;全文 *

Also Published As

Publication number Publication date
CN106303726A (en) 2017-01-04

Similar Documents

Publication Publication Date Title
CN106303726B (en) Video tag adding method and device
US10999634B2 (en) Media content recommendation method, server, client, and storage medium
CN104575553B (en) Produce the method and apparatus for being superimposed upon and playing the barrage on object
CN103079092B (en) Obtain the method and apparatus of people information in video
EP3152948B1 (en) Place-based information processing method and apparatus
WO2017166517A1 (en) Method and device for interaction in live broadcast
KR102370699B1 (en) Method and apparatus for acquiring information based on an image
US20190230311A1 (en) Video interface display method and apparatus
US20140176487A1 (en) Communication terminal, image displaying system, processing method in a communication terminal, and computer program
WO2017101318A1 (en) Method and client for implementing voice interaction in live video broadcast process
CN104376121B (en) System, method and the user terminal that a kind of picture Adaptive matching combination is presented
CN111182359A (en) Video preview method, video frame extraction method, video processing device and storage medium
WO2022184030A1 (en) Wearable device interaction method and apparatus
CN111479119A (en) Method, device and system for collecting feedback information in live broadcast and storage medium
CN104333772A (en) Intelligent notification method for ordering live television program
CN104349212A (en) Method for playing advertisement and video content and electronic equipment for realizing method
CN112688859A (en) Voice message sending method and device, electronic equipment and readable storage medium
JP2018207220A (en) Relative information display system
CN111586492A (en) Video playing method and device, client device and storage medium
CN117857833A (en) Live broadcast processing method, device, terminal and medium
CN114666643A (en) Information display method and device, electronic equipment and storage medium
CN120428905A (en) Multimedia information display method, device, electronic device and storage medium
KR101612000B1 (en) System for purchasing advertisement object based on logotional advertisement using movement of purview cursor area in mobile smart terminal
CN120578825A (en) Page display method, device, equipment and storage medium
CN119922388A (en) A video playing method, a video processing method and a corresponding device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant