WO2025087392A1 - Multimedia data processing method and apparatus, and electronic device and storage medium - Google Patents
Multimedia data processing method and apparatus, and electronic device and storage medium Download PDFInfo
- Publication number
- WO2025087392A1 WO2025087392A1 PCT/CN2024/127406 CN2024127406W WO2025087392A1 WO 2025087392 A1 WO2025087392 A1 WO 2025087392A1 CN 2024127406 W CN2024127406 W CN 2024127406W WO 2025087392 A1 WO2025087392 A1 WO 2025087392A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- multimedia data
- reference image
- image
- original multimedia
- color
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/86—Camera processing pipelines; Components thereof for processing colour signals for controlling the colour saturation of colour signals, e.g. automatic chroma control circuits
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/85—Camera processing pipelines; Components thereof for processing colour signals for matrixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/88—Camera processing pipelines; Components thereof for processing colour signals for colour balance, e.g. white-balance circuits or colour temperature control
Definitions
- the embodiments of the present disclosure relate to image technology, and more particularly to a method, device, electronic device and storage medium for processing multimedia data.
- the current color adjustment method mainly adjusts parameters manually, which makes it difficult to quickly and accurately obtain a color adjustment result that meets the color adjustment requirements.
- an embodiment of the present disclosure provides a method for processing multimedia data, comprising: obtaining original multimedia data; obtaining a reference image; adjusting image frames in the original multimedia data according to display properties of the reference image to obtain target multimedia data, and displaying the target multimedia data.
- the target multimedia data includes a sequence of image frames in the original multimedia data that are adjusted based on the display properties of the reference image.
- the embodiments of the present disclosure further provide a multimedia data processing device, the device comprising: a data acquisition module, used to acquire original multimedia data; an image acquisition module, used to acquire a reference image; an image adjustment module, used to adjust the image frames in the original multimedia data according to the display attributes of the reference image, obtain target multimedia data, and display the target multimedia data.
- the target multimedia data includes a sequence of image frames in the original multimedia data that are adjusted based on the display attributes of the reference image.
- an embodiment of the present disclosure further provides an electronic device, comprising: one or more processors; a storage device for storing one or more programs, wherein when the one or more programs are executed by the one or more processors, the one or more processors implement a method for processing multimedia data as described in any embodiment of the present disclosure.
- an embodiment of the present disclosure further provides a storage medium comprising computer executable instructions, which, when executed by a computer processor, are used to execute the method for processing multimedia data as described in any embodiment of the present disclosure.
- FIG1 is a schematic flow chart of a method for processing multimedia data provided by an embodiment of the present disclosure
- FIG2 is a schematic flow chart of another method for processing multimedia data provided by an embodiment of the present disclosure.
- FIG3 is a schematic diagram of superimposing and displaying a reference image in an image frame of original multimedia data provided by an embodiment of the present disclosure
- FIG4 is a schematic diagram of a reference image superimposed with line material provided by an embodiment of the present disclosure.
- FIG5 is a schematic diagram of a color block special effect provided by an embodiment of the present disclosure.
- FIG6 is a schematic diagram of a color block shift provided by an embodiment of the present disclosure.
- FIG7 is a schematic diagram of the structure of a multimedia data processing device provided by an embodiment of the present disclosure.
- FIG8 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present disclosure.
- the embodiments of the present disclosure provide a method, device, electronic device and storage medium for processing multimedia data, which can improve color adjustment efficiency and accuracy.
- the disclosed embodiments provide a method, device, electronic device and storage medium for processing multimedia data, which adjusts the image frames in the original multimedia data by the display attributes of the reference image, obtains the target multimedia data, and displays the target multimedia data, so as to adjust the display style of the original multimedia data based on the color style of the reference image, so that the target multimedia data presents a display effect similar to the reference image, omits the step of manually adjusting the display attributes, and improves the color adjustment efficiency. Since the reference image can intuitively present the color adjustment result desired by the user, adjusting the original multimedia data based on the display attributes of the reference image can improve the adjustment accuracy and avoid repeated modification of parameters.
- Figure 1 is a flow chart of a method for processing multimedia data provided by an embodiment of the present disclosure.
- the embodiment of the present disclosure is applicable to the case of video color grading.
- the method can be executed by a multimedia data processing device, which can be implemented in the form of software and/or hardware.
- a multimedia data processing device which can be implemented in the form of software and/or hardware.
- an electronic device which can be a mobile terminal, a PC or a server, etc.
- the method comprises:
- the original multimedia data may be a video or image selected by the user to be color-adjusted, etc.
- a historical video in the client may be specified as the original multimedia data.
- multimedia data captured by the user in real time may be obtained as the original multimedia data.
- multimedia data selected by the user from a local album may be obtained as the original multimedia data.
- the original multimedia data is obtained and played. The display properties of the original multimedia data are adjusted without interrupting the playback.
- the reference image represents a reference image for adjusting the display properties of the image frame of the original multimedia data.
- the reference image may be an image uploaded by a user that meets the color adjustment requirements.
- the reference image may be an image taken by a user, or may be an image downloaded from the Internet, and the embodiment of the present disclosure does not limit the source of the reference image.
- the reference image uploaded by the user is obtained. Since the display style of the reference image meets the color adjustment requirements of the user, the step of repeatedly modifying parameters in order to obtain a color adjustment result that meets the color adjustment requirements during the manual color adjustment process can be avoided, thereby improving the color adjustment efficiency.
- the display attribute represents the display effect of the reference image in terms of color and brightness.
- the display attribute may include brightness attribute and color attribute.
- the target multimedia data may be Characterize the color adjustment result of the original multimedia data.
- the target multimedia data includes a sequence of image frames in the original multimedia data adjusted based on the display attributes of the reference image, and the target multimedia data presents a display effect similar to the reference image. Exemplarily, for the image frames in the reference image and the original multimedia data, the target pixel interval corresponding to the image is determined according to the display attributes respectively.
- attribute matching is performed on the image frame in the original multimedia data based on the display attributes of the pixels of the reference image.
- the attribute matching results corresponding to each target pixel interval are fused to obtain the multimedia data processing results corresponding to the image frame.
- Attribute matching may include matching attributes such as color and brightness.
- the attribute matching result represents the processing result of migrating the attribute features of the reference image to the image frame of the original multimedia data. By migrating the attribute features of the reference image to the image frame of the original multimedia data, the image frame of the original multimedia data presents display attributes similar to the reference image.
- the display attributes of the pixel are adjusted according to the display attributes of the reference image to obtain the skin color adjustment result corresponding to the image frame.
- the skin color adjustment result and the multimedia data processing result are fused according to the skin color weight to obtain the target multimedia data.
- the skin color weight represents the probability that each pixel in the image frame of the original multimedia data belongs to the skin color area.
- the reference image and the image frame of the original multimedia data are respectively converted into color space. Based on the color space conversion result, the brightness channel data of the reference image and the image frame of the original multimedia data are respectively obtained. According to the brightness channel data, the histograms corresponding to the reference image and the image frame of the original multimedia data are respectively determined. According to the number of pixels corresponding to each brightness value in the histogram of the reference image, the brightness segmentation point corresponding to the reference image is determined. Then, according to the brightness segmentation point corresponding to the reference image, the reference image is segmented into a plurality of target pixel intervals.
- the brightness segmentation point corresponding to the image frame in the original multimedia data is determined. Then, according to the brightness segmentation point corresponding to the image frame, the image frame is segmented into a plurality of target pixel intervals.
- the target pixel intervals include shadow intervals, midtone intervals, and highlight intervals.
- the pixels of the reference image are counted according to the brightness attribute, and the brightness segmentation points are determined according to the pixel statistics. Then, the reference image is segmented into a plurality of target pixel intervals according to the brightness segmentation points. The number of target pixel intervals is determined based on the number of brightness segmentation points, and adjacent target pixel intervals have overlapping areas.
- the reference image is converted to a YUV color space.
- the Y channel data is determined to be a brightness attribute, and a histogram of the Y channel of the reference image is generated through the Y channel data.
- the pixels of the shadow, midtone, and highlight are determined according to each brightness value in the histogram of the Y channel.
- the alternative segmentation point is determined so that the number of pixels corresponding to the shadow, midtone, and highlight is the same.
- the alternative segmentation point is determined so that the areas of the shadow interval, the midtone interval, and the highlight interval are all the same.
- adjacent intervals can be staggered by a certain range so that adjacent intervals have overlapping areas.
- the area of the overlapping area can be set according to the actual application scenario. For example, the overlapping area can be set to n times the area of the color area corresponding to the reference image, and 0 ⁇ n ⁇ 1.
- the brightness segmentation point is determined based on the brightness value corresponding to the starting point and the brightness value corresponding to the end point of each overlapping area. For example, if the target pixel interval includes the shadow interval, the midtone interval and the highlight interval, there are two overlapping areas in the target pixel interval, namely the overlapping area between the shadow interval and the midtone interval and the overlapping area between the midtone interval and the highlight interval. There are four brightness segmentation points in the target pixel interval, namely S1, S2, S3 and S4, and S1 ⁇ S2 ⁇ S3 ⁇ S4.
- the range of the shadow interval is [0, S2]
- the range of the midtone interval is [S1, S4]
- the range of the highlight interval is [S3, 255].
- the brightness segmentation points and the range of the target pixel interval of the image frame in the original multimedia data are determined in a manner similar to the above manner, which will not be described in detail here.
- the first weight corresponding to the pixel and each target pixel interval is determined according to the brightness attribute of the pixel and the brightness segmentation point corresponding to the overlapping area.
- the first weight corresponding to the pixel and each target pixel interval is determined according to the target pixel interval to which the pixel belongs.
- a number of first weights can be determined according to its brightness.
- the number of first weights is consistent with the number of target pixel intervals. For example, if the image frame in the original multimedia data is divided into a shadow interval, a midtone interval, and a highlight interval, three first weights of each pixel in the image frame in the original multimedia data can be calculated. That is, the shadow weight W shadow of the pixel corresponding to the shadow interval, the midtone weight W mid of the pixel corresponding to the midtone interval, and the highlight weight W highlight of the pixel corresponding to the highlight interval, and the sum of the three weights is 1.
- the first weight of the pixel corresponding to the highlight range 0.
- the first weights of pixels completely in the midtone range and the highlight range can be determined in the same manner, which will not be repeated here.
- the pixel has both a shadow weight and a midtone weight.
- the first weight W shadow of the pixel corresponding to the shadow interval is determined
- the first weight W mid of the pixel corresponding to the midtone interval is determined
- the first weight W highlight of the pixel corresponding to the highlight interval is determined to be 0.
- the first weight corresponding to the reference image can be determined by using a method similar to the above, which will not be described in detail here.
- attribute matching is performed on the image frames in the original multimedia data based on the display attributes of the pixels of the reference image.
- the attribute matching results corresponding to each of the target pixel intervals are fused to obtain the multimedia data processing results corresponding to the image frames, including: performing brightness matching on the image frames in the original multimedia data based on the brightness attribute of the reference image to obtain brightness matching results of the image frames in the original multimedia data.
- the image frames in the original multimedia data are converted to the YUV color space. Color matching is performed on the UV channel, and brightness matching is performed on the Y channel.
- a histogram matching algorithm is used to perform histogram matching on the Y channel of the image frame in the original multimedia data and the Y channel of the reference image. Match and get the brightness matching result.
- the color features and transformation matrix corresponding to the target pixel interval are determined according to the color attributes of each target pixel interval.
- the color features represent the color distribution of pixels in the target pixel interval corresponding to the image frame and the reference image in the original multimedia data.
- the color features may include UV channel mean and covariance matrix, etc.
- the UV channel data of the pixels in the shadow interval, the midtone interval and the highlight interval are obtained respectively.
- the UV mean and the covariance matrix are calculated according to the UV channel data.
- the transformation matrix M is determined according to the UV mean and the covariance matrix.
- the color features and transformation matrix corresponding to the target pixel interval are determined according to the color attributes of each target pixel interval, including: determining the first color mean and the first covariance matrix corresponding to the target pixel interval according to the color attributes of the pixels in each target pixel interval of the image frame in the original multimedia data. Determining the second color mean and the second covariance matrix corresponding to the target pixel interval according to the color attributes of the pixels in each target pixel interval of the reference image; determining the transformation matrix corresponding to the target pixel interval according to the first color mean, the second color mean, the first covariance matrix and the second covariance matrix, wherein the color features include the first color mean and the second color mean.
- the UV channel mean of all shadow pixels in the image frame in the original multimedia data is calculated.
- Covariance matrix ⁇ src The first color mean can be expressed as and The first covariance matrix can be expressed as ⁇ src .
- Calculate the UV channel mean of all shadow pixels in the reference image The covariance matrix ⁇ ref .
- the second color mean can be expressed as and The second covariance matrix can be expressed as ⁇ ref .
- the transformation matrix M is determined based on the above mean and covariance matrix using the MKL algorithm.
- the method further includes: determining an eigenvalue of the first covariance matrix.
- the eigenvalue represents the shape of an ellipse presented by the pixels of the image frame in the original multimedia data according to the Gaussian distribution.
- the unit matrix is used as the transformation matrix.
- the preset condition represents the shape threshold of the ellipse presented by the Gaussian distribution of the pixels of the image frame in the original multimedia data. If the Gaussian distribution exceeding the shape threshold is subjected to color transformation, color abnormal points may be generated. In this case, the transformation matrix is calculated based on the covariance matrix, and the transformation matrix M is set to the unit matrix instead.
- the eigenvalues e1 and e2 of the covariance matrix are determined according to the covariance matrix ⁇ src of the image frame in the original multimedia data, and e1>e2.
- the eigenvalues of the covariance matrix represent the major axis e1 and minor axis e2 of the ellipse presented by the pixels of the image frame in the original multimedia data according to the Gaussian distribution. If e1/e2> the set shape threshold, it means that the ellipse presented by the Gaussian distribution is very flat. At this time, if the Gaussian distribution is subjected to color transformation, color outliers are likely to be generated. Therefore, the transformation matrix M is set to the unit matrix.
- the first color mean values of the image frames in the original multimedia data in the shadow interval, the mid-tone interval and the highlight interval can be calculated respectively, which are: and and and
- the covariance matrices of the image frames in the original multimedia data in the shadow interval, the midtone interval and the highlight interval are calculated respectively, namely ⁇ src0 , ⁇ src1 , ⁇ src2 .
- the transformation matrices M 0 , M 1 , and M 2 of the image frames in the original multimedia data in the shadow interval, the midtone interval and the highlight interval are calculated respectively.
- Color matching is performed on each of the target pixel intervals corresponding to the image frame in the original multimedia data according to the color features and the transformation matrix.
- the second weight corresponding to the pixel is determined according to color information of the pixel in the target pixel interval and the first color mean value.
- the second weight represents the degree of deviation of the color of the pixel from the mean in the Gaussian distribution.
- the smaller the second weight the more the color value of the pixel deviates from the mean of the Gaussian distribution.
- the larger the second weight the closer the color value of the pixel is to the mean of the Gaussian distribution.
- the color information uv of the pixels in the shadow interval of the image frame in the original multimedia data in the UV channel can be expressed as a two-dimensional vector
- the first color mean value of the shadow interval of the image frame in the original multimedia data It can be expressed as According to the pixel color information uv, the first color mean
- the second weight is calculated by using the first covariance matrix ⁇ src0 .
- the second weights corresponding to the pixels in the midtone interval and the highlight interval are calculated, which will not be repeated here.
- the shadow interval Take the shadow interval as an example. Perform a color transformation on each pixel in the shadow interval of the image frame in the original multimedia data based on the transformation matrix to obtain new color information as the first color matching result. Adjust the color information of each pixel in the shadow interval of the image frame in the original multimedia data based on the color mean of the image frame in the original multimedia data and the reference image in the shadow interval to obtain new color information as the second color matching result. For example, the color information uv of the pixel of the image frame in the original multimedia data is combined with the transformation matrix M 0 corresponding to the shadow interval to obtain the first color matching result corresponding to the shadow interval, thereby realizing the color style transfer of the reference image to the image frame in the original multimedia data.
- the color information uv of the pixel of the image frame in the original multimedia data minus the color information uv of the pixel of the image frame in the original multimedia data Back Fold With reference image A second color matching result corresponding to the shadow interval is obtained.
- the second color mean of the shadow interval of the reference image can be expressed as
- the first color matching result and the second color matching result are fused according to the second weight to obtain a color matching result of the target pixel interval.
- the fusion coefficient is determined according to the second weight, and the first color matching result and the second color matching result corresponding to the shadow interval are fused according to the fusion coefficient to obtain the color matching result of the shadow interval.
- the fusion coefficient includes the second weight, and the sum of the fusion coefficients is 1.
- the color matching results of the midtone interval and the color matching results of the highlight interval are calculated in a similar manner, which will not be repeated here.
- the midtone interval or the highlight interval there is only a color matching result for the shadow interval.
- the midtone interval For pixels in the image frame in the original multimedia data whose brightness belongs to the overlapping area of the shadow interval and the midtone interval, there are a color matching result for the shadow interval and a color matching result for the midtone interval.
- the midtone interval For pixels in the image frame in the original multimedia data whose brightness belongs to the overlapping area of the midtone interval and the highlight interval, there are a color matching result for the midtone interval and a color matching result for the highlight interval.
- the color matching result of the target pixel interval is fused according to the first weight corresponding to the pixel and each target pixel interval to obtain the color matching result of the image frame in the original multimedia data; and the processing result image of the image frame in the original multimedia data is determined according to the color matching result and brightness matching result of the image frame in the original multimedia data as the multimedia data processing result corresponding to the image frame.
- the first weights i.e., shadow weight, midtone weight, and highlight weight
- the respective target pixel intervals i.e., shadow interval, midtone interval, and highlight interval
- the color matching results of the pixel in the corresponding target pixel intervals i.e., shadow weight, midtone weight, and highlight weight
- the color matching result of the pixel in the shadow interval For example, for each pixel of the image frame in the original multimedia data, according to the shadow weight, midtone weight and highlight weight of the pixel, the color matching result of the pixel in the shadow interval, the color matching result of the pixel in the midtone interval and the color matching result of the pixel in the highlight interval are fused. Thus, the color matching result of the image frame in the original multimedia data is obtained.
- the display attributes of the pixels are adjusted according to the display attributes of the skin color area of the reference image to obtain a skin color adjustment result corresponding to the image frame. This includes: determining the probability that each pixel in the image frame of the original multimedia data belongs to a skin color area as a skin color weight, and determining the pixels included in the skin color area in the image frame of the original multimedia data according to the skin color weight.
- the display attributes of the pixels are adjusted according to the mean values of the display attributes of the image frame and the reference image in the skin color area, respectively, to obtain a skin color adjustment result corresponding to the image frame.
- the original multimedia data is input into the skin color segmentation model.
- the skin color segmentation model is used to perform skin color segmentation on the image frames in the original multimedia data, and the probability that each pixel in the image frame belongs to the skin color area is obtained. The pixels contained in the skin color area of the image frame in the original multimedia data are determined according to the probability.
- the reference image is input into the skin color segmentation model.
- the skin color segmentation model is used to perform skin color segmentation on the reference image, and the probability that each pixel in the reference image belongs to the skin color area is obtained. The pixels contained in the skin color area of the reference image are determined according to the probability.
- the brightness mean of the skin color area of the image frame in the original multimedia data is determined.
- the brightness mean of the skin color area of the reference image is determined.
- the chromaticity of each pixel contained in the skin color area of the image frame in the original multimedia data is determined.
- the chromaticity mean of the skin color area of the reference image is determined.
- the difference between the brightness of the pixel and the average brightness of the skin color area is calculated, and the difference is superimposed with the average brightness of the skin color area of the reference image to obtain a brightness adjustment result of the pixel.
- the difference between the chromaticity of the pixel and the chromaticity mean of the skin color area is calculated, and the difference is superimposed with the chromaticity mean of the skin color area of the reference image to obtain the chromaticity adjustment result of the pixel.
- a skin color adjustment result corresponding to the image frame is obtained.
- the target multimedia data is obtained by fusing the multimedia data processing result and the skin color adjustment result of the image frame in a linear fusion manner based on the skin color weight. For example, the target multimedia data is determined based on the fusion result of the multimedia data processing result and the skin color adjustment result corresponding to the image frame sequence included in the target multimedia data.
- the technical solution of the disclosed embodiment adjusts the image frame in the original multimedia data by the display attribute of the reference image, obtains the target multimedia data and displays it.
- the display style of the original multimedia data is adjusted based on the color style of the reference image, so that the target multimedia data presents a display effect similar to that of the reference image, omitting the step of manually adjusting the display attribute, and improving the color adjustment efficiency. Since the reference image can intuitively present the color adjustment result desired by the user, adjusting the original multimedia data based on the display attribute of the reference image can improve the adjustment accuracy and avoid repeated modification of parameters.
- FIG. 2 is a flow chart of another method for processing multimedia data provided by an embodiment of the present disclosure. Embodiment Based on the above embodiment, additional display steps of reference images and special effect elements are added. As shown in FIG2 , the method includes: S210, obtaining original multimedia data. S220, obtaining a reference image. S230, generating special effect elements according to the reference image, and superimposing and displaying the reference image and special effect elements on the original multimedia data.
- the display effect is relatively simple and the user experience is poor. It may even cause the user to mistakenly believe that the reference image has not been successfully uploaded and repeat the operation, resulting in erroneous processing.
- the single display effect and the erroneous operation of repeatedly uploading the reference image can be avoided.
- the special effect element is superimposed on the set position of the reference image and is displayed and moves in a set direction.
- the special effect element includes a color block special effect, etc.
- the color block special effect may include color blocks of different colors.
- the color of the color block represents the color category obtained by clustering the colors of the pixels of the reference image.
- the color of the color block may also be a fixed color, and the embodiment of the present disclosure does not specifically limit the shape and color of the color block.
- the color category included in the reference image is determined according to the color attribute of the pixels in the reference image, and a special effect element corresponding to the color represented by the color category is generated.
- multiple cluster centers are pre-set, and pixels in the reference image are clustered according to the colors of the pixels and the colors of the pixels corresponding to the cluster centers to obtain multiple pixel groups.
- the colors corresponding to the cluster centers of all groups of pixels can be used as the color categories included in the reference image.
- FIG3 is a schematic diagram of superimposing and displaying a reference image in an image frame of original multimedia data provided by an embodiment of the present disclosure. As shown in FIG3, during the playback of the original multimedia data, the reference image is superimposed and displayed at a first position 310 of the original multimedia data at time t1 , and slides to a second position 320 at time t2 .
- FIG4 is a schematic diagram of a reference image superimposed with line material provided by an embodiment of the present disclosure.
- a mask image 410 with set transparency is generated according to the width and height of the reference image, and there are bright areas 411 and dark areas 412 in the mask image 410.
- a dynamic image is synthesized based on multiple mask images 410 with different positions of bright areas 411. The dynamic image is superimposed on the reference image to obtain the effect of light scanning the reference image.
- the special effect element is superimposed on the first area of the reference image according to the set superposition form.
- the set superposition form represents the form of superimposing the color block special effect on the reference image.
- the set superposition form can be to display the color blocks in different positions of the reference image in sequence according to the color of the color blocks.
- the first area can be a preset area within the reference image.
- FIG5 is a schematic diagram of a color block special effect provided by an embodiment of the present disclosure. As shown in FIG5, each color block 520 is displayed one by one in the reference image 510 until all the color blocks 520 are displayed in the first area of the reference image 510. During the display process, the width and height of the color block are changed from a first state to a second state.
- the display position and brightness arrangement order corresponding to the special effect element are determined according to the set superposition form.
- the display order of the special effect element is determined according to the brightness arrangement order.
- the special effect element is superimposed and displayed in the first area of the reference image in the image frame in the original multimedia data.
- the display position represents the position of the special effect element in the reference image.
- the display position may include color blocks arranged horizontally (or vertically or obliquely) in the reference image, and the spacing between adjacent color blocks is the same (or gradually increases or gradually decreases).
- the brightness arrangement order may include color or brightness arranged from dark to light, or from light to dark, etc.
- each color block represents the color appearing in the reference image, and the color blocks have different color depths and different brightness.
- the display order of the color blocks in the reference image can be determined according to the brightness sorting order specified in the set superposition form. Then, the color blocks are displayed one by one at the display position corresponding to the color blocks according to the display order.
- the first state of the color block is first displayed in the reference image, and the process of the color block converting from the first state to the second state is presented. For example, a small color block is first displayed in the reference image, and then the small color block is enlarged to obtain a color block with a set width and height.
- FIG. 6 is a schematic diagram of a color block shift provided by an embodiment of the present disclosure. As shown in Figure 6, after all color blocks are displayed in the first area of the reference image, all color blocks 610 are rotated from a horizontal arrangement to a vertical arrangement, and moved to the second area 620 corresponding to the reference image. After all color blocks 610 reach the second area 620, the spacing between adjacent color blocks is increased. S240. Adjust the image frame in the original multimedia data according to the display attributes of the reference image to obtain target multimedia data. S250. When the display state of the special effect element in the original multimedia data meets the set conditions, display the target multimedia data.
- the set condition may represent a condition under which the special effect element disappears from the original multimedia data, and when the special effect element disappears from the original multimedia data, the target multimedia data is displayed. For example, when the spacing between adjacent color blocks is increased so that the spacing between adjacent color blocks is a set value, it is determined that the special effect color meets the set condition. Alternatively, when the spacing between adjacent color blocks is increased and the display time of the color blocks reaches a set time, it is determined that the special effect color meets the set condition.
- the technical solution of the disclosed embodiment can intuitively display the color style of the reference image selected by the user by generating special effect elements corresponding to the reference image and superimposing the reference image and the special effect elements in the original multimedia data. It can also avoid erroneous operations such as a single display effect or repeated uploading of reference images in the process of adjusting the display properties of the original multimedia data based on the reference image by vividly displaying the special effect elements, thereby increasing interactivity and fun and improving user experience.
- FIG. 7 is a schematic diagram of the structure of a multimedia data processing device provided in an embodiment of the present disclosure.
- the device may be implemented in the form of software and/or hardware.
- it may be implemented by an electronic device, which may be a mobile terminal, a PC, a server, etc.
- the device includes: an image acquisition module 710 and an image adjustment module 720.
- Block 710 is used to obtain original multimedia data;
- image acquisition module 720 is used to obtain a reference image;
- image adjustment module 730 is used to adjust the image frames in the original multimedia data according to the display properties of the reference image to obtain target multimedia data and display the target multimedia data.
- the target multimedia data includes a sequence of image frames in the original multimedia data adjusted based on the display properties of the reference image.
- the method further includes: a special effect element generation module, which is used to generate a special effect element according to the reference image after acquiring the reference image, and to superimpose and display the reference image and the special effect element in the original multimedia data.
- the special effect element is superimposed on the reference image and displayed at a set position and moves in a set direction; when the display state of the special effect element in the original multimedia data meets the set condition, the step of displaying the target multimedia data is executed.
- generating a special effect element according to the reference image includes: determining a color category included in the reference image according to color attributes of pixels in the reference image; and generating a special effect element corresponding to a color represented by the color category.
- the superimposing and displaying the reference image and special effect element in the original multimedia data includes: superimposing and displaying the reference image in the image frame of the original multimedia data; superimposing and displaying line material in the reference image, wherein the line material moves along a first direction in the reference image; superimposing the special effect element to a first area in the reference image according to a set superimposition form; moving the special effect element from the first area of the reference image along a second direction to a second area corresponding to the reference image, and adjusting the display status of the special effect element in the second area.
- superimposing the special effect element to the first area in the reference image according to the set superimposition form includes: determining a display position and a brightness arrangement order corresponding to the special effect element according to the set superimposition form; determining a display order of the special effect elements according to the brightness arrangement order; and superimposing and displaying the special effect element in the first area of the reference image in the image frame of the original multimedia data according to the display position and display order.
- the image adjustment module 720 is specifically used to: for the reference image and the image frames in the original multimedia data, determine the target pixel intervals corresponding to the images according to the display attributes respectively; for each of the target pixel intervals, perform attribute matching on the image frames in the original multimedia data based on the display attributes of the pixels of the reference image, fuse the attribute matching results corresponding to each of the target pixel intervals, and obtain the multimedia data processing results corresponding to the image frames; for the pixels in the skin color area of the image frames of the original multimedia data, adjust the display attributes of the pixels according to the display attributes of the skin color area of the reference image, and obtain the skin color adjustment results corresponding to the image frames; for the image frames in the original multimedia data, fuse the skin color adjustment results and the multimedia data processing results according to the skin color weight, and obtain the target multimedia data.
- the skin color weight represents the probability that each pixel in the image frame of the original multimedia data belongs to the skin color area.
- adjusting the display attribute of the pixel according to the display attribute of the skin color area of the reference image to obtain a skin color adjustment result corresponding to the image frame includes: determining a probability that each pixel in the image frame of the original multimedia data belongs to the skin color area, as the skin color
- the method comprises the steps of: determining pixels included in a skin color area in an image frame of the original multimedia data according to the skin color weight; determining a mean value of display attributes of the skin color area in the image frame of the original multimedia data according to display attributes of the pixels included in the skin color area in the image frame of the original multimedia data; and adjusting the display attributes of the pixels in the skin color area in the image frame of the original multimedia data according to the mean values of display attributes of the image frame and a reference image in the skin color area, so as to obtain a skin color adjustment result corresponding to the image frame.
- the multimedia data processing device provided in the embodiments of the present disclosure can execute the multimedia data processing method provided in any embodiment of the present disclosure, and has the corresponding functional modules and beneficial effects of the execution method.
- FIG8 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present disclosure.
- the terminal device in the embodiment of the present disclosure may include, but is not limited to, mobile terminals such as mobile phones, laptop computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), vehicle-mounted terminals (e.g., vehicle-mounted navigation terminals), etc., and fixed terminals such as digital TVs, desktop computers, etc.
- the electronic device shown in FIG8 is merely an example and should not impose any limitations on the functions and scope of use of the embodiments of the present disclosure.
- the electronic device 800 may include a processing device (e.g., a central processing unit, a graphics processing unit, etc.) 801, which can perform various appropriate actions and processes according to a program stored in a read-only memory (ROM) 802 or a program loaded from a storage device 808 to a random access memory (RAM) 803.
- a processing device 801 e.g., a central processing unit, a graphics processing unit, etc.
- RAM random access memory
- various programs and data required for the operation of the electronic device 800 are also stored.
- the processing device 801, the ROM 802, and the RAM 803 are connected to each other via a bus 804.
- An edit/output (I/O) interface 805 is also connected to the bus 804.
- the following devices may be connected to the I/O interface 805: input devices 806 including, for example, a touch screen, a touchpad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; output devices 807 including, for example, a liquid crystal display (LCD), a speaker, a vibrator, etc.; storage devices 808 including, for example, a magnetic tape, a hard disk, etc.; and communication devices 809.
- the communication device 809 may allow the electronic device 800 to communicate wirelessly or wired with other devices to exchange data.
- FIG. 8 shows an electronic device 800 with various devices, it should be understood that it is not required to implement or have all the devices shown. More or fewer devices may be implemented or have alternatively.
- an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a non-transitory computer-readable medium, and the computer program contains program code for executing the method shown in the flowchart.
- the computer program can be downloaded and installed from a network through a communication device 809, or installed from a storage device 808, or installed from a ROM 802.
- the processing device 801 the execution The above functions are defined in the method of the embodiment of the present disclosure.
- the electronic device provided by the embodiment of the present disclosure and the method for processing multimedia data provided by the above embodiment belong to the same inventive concept.
- the embodiments of the present disclosure provide a computer storage medium on which a computer program is stored.
- the program is executed by a processor, the method for processing multimedia data provided by the above embodiments is implemented.
- the computer-readable medium disclosed above may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two.
- the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination of the above.
- Computer-readable storage media may include, but are not limited to: an electrical connection with one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above.
- a computer-readable storage medium may be any tangible medium containing or storing a program that may be used by or in combination with an instruction execution system, device or device.
- a computer-readable signal medium may include a data signal propagated in a baseband or as part of a carrier wave, in which a computer-readable program code is carried.
- This propagated data signal may take a variety of forms, including but not limited to an electromagnetic signal, an optical signal, or any suitable combination of the above.
- the computer readable signal medium may also be any computer readable medium other than a computer readable storage medium, which may send, propagate or transmit a program for use by or in conjunction with an instruction execution system, apparatus or device.
- the program code contained on the computer readable medium may be transmitted using any suitable medium, including but not limited to: wires, optical cables, RF (radio frequency), etc., or any suitable combination of the above.
- the client and server may communicate using any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol), and may be interconnected with any form or medium of digital data communication (e.g., a communication network).
- HTTP HyperText Transfer Protocol
- Examples of communication networks include a local area network ("LAN”), a wide area network ("WAN”), an internet (e.g., the Internet), and a peer-to-peer network (e.g., an ad hoc peer-to-peer network), as well as any currently known or future developed network.
- the computer-readable medium may be included in the electronic device, or may exist independently without being incorporated into the electronic device.
- the computer-readable medium carries one or more programs.
- the electronic device When the one or more programs are executed by the electronic device, the electronic device: obtains original multimedia data; obtains a reference image; adjusts the image frame in the original multimedia data according to the display attribute of the reference image to obtain target multimedia data, and displays the target multimedia data.
- the target multimedia data includes an image frame sequence in the original multimedia data that is adjusted based on the display attributes of the reference image.
- Computer program code for performing the operations of the present disclosure may be written in one or more programming languages or a combination thereof, including, but not limited to, object-oriented programming languages, such as Java, Smalltalk, C++, and conventional procedural programming languages, such as "C" or similar programming languages.
- the program code may be executed entirely on the user's computer, partially on the user's computer, as a separate software package, partially on the user's computer and partially on a remote computer, or entirely on a remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (e.g., through the Internet using an Internet service provider).
- LAN local area network
- WAN wide area network
- Internet service provider e.g., AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
- each square box in the flow chart or block diagram can represent a module, a program segment or a part of a code, and the module, the program segment or a part of the code contains one or more executable instructions for realizing the specified logical function.
- the functions marked in the square box can also occur in a sequence different from that marked in the accompanying drawings. For example, two square boxes represented in succession can actually be executed substantially in parallel, and they can sometimes be executed in the opposite order, depending on the functions involved.
- each square box in the block diagram and/or flow chart, and the combination of the square boxes in the block diagram and/or flow chart can be implemented with a dedicated hardware-based system that performs a specified function or operation, or can be implemented with a combination of dedicated hardware and computer instructions.
- the units involved in the embodiments described in the present disclosure may be implemented by software or hardware.
- the name of a unit does not limit the unit itself in some cases.
- exemplary types of hardware logic components include: field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on chips (SOCs), complex programmable logic devices (CPLDs), and the like.
- FPGAs field programmable gate arrays
- ASICs application specific integrated circuits
- ASSPs application specific standard products
- SOCs systems on chips
- CPLDs complex programmable logic devices
- a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, device, or equipment.
- a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
- a machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or equipment, or any suitable combination of the foregoing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
Abstract
Description
本申请要求2023年10月27日递交的、标题为“一种多媒体数据的处理方法、装置、电子设备及存储介质”、申请号为202311415449X的中国发明专利申请的优先权,该申请的全部内容通过引用结合在本申请中。This application claims priority to the Chinese invention patent application entitled “A method, device, electronic device and storage medium for processing multimedia data” filed on October 27, 2023, with application number 202311415449X. The entire contents of that application are incorporated by reference into this application.
本公开实施例涉及图像技术,尤其涉及一种多媒体数据的处理方法、装置、电子设备及存储介质。The embodiments of the present disclosure relate to image technology, and more particularly to a method, device, electronic device and storage medium for processing multimedia data.
随着移动终端以及互联网的普及应用,越来越多的用户可以拍摄多媒体数据,并通过互联网分享多媒体数据。With the popularization and application of mobile terminals and the Internet, more and more users can capture multimedia data and share the multimedia data through the Internet.
在分享之前,用户可能需要对多媒体数据的色彩风格进行调整,以得到一个满足要求的调色结果。目前的调色方式主要通过手动设置参数调整,难以快速、准确地得到满足调色要求的调色结果。Before sharing, users may need to adjust the color style of multimedia data to obtain a color adjustment result that meets the requirements. The current color adjustment method mainly adjusts parameters manually, which makes it difficult to quickly and accurately obtain a color adjustment result that meets the color adjustment requirements.
发明内容Summary of the invention
第一方面,本公开实施例提供了一种多媒体数据的处理方法,包括:获取原始多媒体数据;获取参考图像;根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据,显示所述目标多媒体数据。所述目标多媒体数据包括所述原始多媒体数据中基于所述参考图像的显示属性调整后的图像帧序列。In a first aspect, an embodiment of the present disclosure provides a method for processing multimedia data, comprising: obtaining original multimedia data; obtaining a reference image; adjusting image frames in the original multimedia data according to display properties of the reference image to obtain target multimedia data, and displaying the target multimedia data. The target multimedia data includes a sequence of image frames in the original multimedia data that are adjusted based on the display properties of the reference image.
第二方面,本公开实施例还提供了一种多媒体数据的处理装置,该装置包括:数据获取模块,用于获取原始多媒体数据;图像获取模块,用于获取参考图像;图像调整模块,用于根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据,显示所述目标多媒体数据。所述目标多媒体数据包括所述原始多媒体数据中基于所述参考图像的显示属性调整后的图像帧序列。In a second aspect, the embodiments of the present disclosure further provide a multimedia data processing device, the device comprising: a data acquisition module, used to acquire original multimedia data; an image acquisition module, used to acquire a reference image; an image adjustment module, used to adjust the image frames in the original multimedia data according to the display attributes of the reference image, obtain target multimedia data, and display the target multimedia data. The target multimedia data includes a sequence of image frames in the original multimedia data that are adjusted based on the display attributes of the reference image.
第三方面,本公开实施例还提供了一种电子设备,所述电子设备包括:一个或多个处理器;存储装置,用于存储一个或多个程序,当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如本公开任意实施例所述的多媒体数据的处理方法。 In a third aspect, an embodiment of the present disclosure further provides an electronic device, comprising: one or more processors; a storage device for storing one or more programs, wherein when the one or more programs are executed by the one or more processors, the one or more processors implement a method for processing multimedia data as described in any embodiment of the present disclosure.
第四方面,本公开实施例还提供了一种包含计算机可执行指令的存储介质,所述计算机可执行指令在由计算机处理器执行时用于执行如本公开任意实施例所述的多媒体数据的处理方法。In a fourth aspect, an embodiment of the present disclosure further provides a storage medium comprising computer executable instructions, which, when executed by a computer processor, are used to execute the method for processing multimedia data as described in any embodiment of the present disclosure.
结合附图并参考以下具体实施方式,本公开各实施例的上述和其他特征、优点及方面将变得更加明显。贯穿附图中,相同或相似的附图标记表示相同或相似的元素。应当理解附图是示意性的,原件和元素不一定按照比例绘制。The above and other features, advantages and aspects of the embodiments of the present disclosure will become more apparent with reference to the following detailed description in conjunction with the accompanying drawings. Throughout the accompanying drawings, the same or similar reference numerals represent the same or similar elements. It should be understood that the drawings are schematic and the originals and elements are not necessarily drawn to scale.
图1为本公开实施例所提供的一种多媒体数据的处理方法的流程示意图;FIG1 is a schematic flow chart of a method for processing multimedia data provided by an embodiment of the present disclosure;
图2为本公开实施例所提供的另一种多媒体数据的处理方法的流程示意图;FIG2 is a schematic flow chart of another method for processing multimedia data provided by an embodiment of the present disclosure;
图3为本公开实施例提供的一种原始多媒体数据的图像帧中叠加显示参考图像的示意图;FIG3 is a schematic diagram of superimposing and displaying a reference image in an image frame of original multimedia data provided by an embodiment of the present disclosure;
图4为本公开实施例提供的一种参考图像叠加线条素材的示意图;FIG4 is a schematic diagram of a reference image superimposed with line material provided by an embodiment of the present disclosure;
图5为本公开实施例提供的一种色块特效的示意图;FIG5 is a schematic diagram of a color block special effect provided by an embodiment of the present disclosure;
图6为本公开实施例提供的一种色块移位的示意图;FIG6 is a schematic diagram of a color block shift provided by an embodiment of the present disclosure;
图7为本公开实施例所提供的一种多媒体数据的处理装置结构示意图;FIG7 is a schematic diagram of the structure of a multimedia data processing device provided by an embodiment of the present disclosure;
图8为本公开实施例所提供的一种电子设备的结构示意图。FIG8 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present disclosure.
下面将参照附图更详细地描述本公开的实施例。虽然附图中显示了本公开的某些实施例,然而应当理解的是,本公开可以通过各种形式来实现,而且不应该被解释为限于这里阐述的实施例,相反提供这些实施例是为了更加透彻和完整地理解本公开。应当理解的是,本公开的附图及实施例仅用于示例性作用,并非用于限制本公开的保护范围。Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. Although certain embodiments of the present disclosure are shown in the accompanying drawings, it should be understood that the present disclosure can be implemented in various forms and should not be construed as being limited to the embodiments described herein, which are instead provided for a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are only for exemplary purposes and are not intended to limit the scope of protection of the present disclosure.
应当理解,本公开的方法实施方式中记载的各个步骤可以按照不同的顺序执行,和/或并行执行。此外,方法实施方式可以包括附加的步骤和/或省略执行示出的步骤。本公开的范围在此方面不受限制。It should be understood that the various steps described in the method embodiments of the present disclosure may be performed in different orders and/or in parallel. In addition, the method embodiments may include additional steps and/or omit the steps shown. The scope of the present disclosure is not limited in this respect.
本文使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”;术语“另一实施例”表示“至少一个另外的实施例”;术语“一些实施例”表示“至少一些实施例”。其他术语的相关定义将在下文描述中给出。The term "including" and its variations used herein are open inclusions, i.e., "including but not limited to". The term "based on" means "based at least in part on". The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments". The relevant definitions of other terms will be given in the following description.
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对不同的装置、模块或单元进行区分,并非用于限定这些装置、模块或单元所执行的功能的顺序或者相互依存关系。It should be noted that the concepts such as "first" and "second" mentioned in the present disclosure are only used to distinguish different devices, modules or units, and are not used to limit the order or interdependence of the functions performed by these devices, modules or units.
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技 术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。It should be noted that the modifications of "one" and "multiple" mentioned in the present disclosure are illustrative rather than restrictive. It should be understood by those skilled in the art that unless the context clearly indicates otherwise, it should be understood as "one or more".
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的,而并不是用于对这些消息或信息的范围进行限制。The names of the messages or information exchanged between multiple devices in the embodiments of the present disclosure are only used for illustrative purposes and are not used to limit the scope of these messages or information.
可以理解的是,本技术方案所涉及的数据(包括但不限于数据本身、数据的获取或使用)应当遵循相应法律法规及相关规定的要求。It is understandable that the data involved in this technical solution (including but not limited to the data itself, the acquisition or use of the data) shall comply with the requirements of relevant laws, regulations and relevant provisions.
本公开实施例提供一种多媒体数据的处理方法、装置、电子设备及存储介质,可以提升调色效率和准确率。The embodiments of the present disclosure provide a method, device, electronic device and storage medium for processing multimedia data, which can improve color adjustment efficiency and accuracy.
本公开实施例提供了一种多媒体数据的处理方法、装置、电子设备及存储介质,通过参考图像的显示属性调整原始多媒体数据中的图像帧,得到目标多媒体数据,并显示目标多媒体数据,实现了基于参考图像的色彩风格调整原始多媒体数据的显示风格,使目标多媒体数据呈现与参考图像相似的显示效果,省略了手动调整显示属性的步骤,提高了调色效率。由于参考图像能够直观的呈现用户想要的调色结果,因此基于参考图像的显示属性调整原始多媒体数据可以提高调整准确率,避免反复修改参数。The disclosed embodiments provide a method, device, electronic device and storage medium for processing multimedia data, which adjusts the image frames in the original multimedia data by the display attributes of the reference image, obtains the target multimedia data, and displays the target multimedia data, so as to adjust the display style of the original multimedia data based on the color style of the reference image, so that the target multimedia data presents a display effect similar to the reference image, omits the step of manually adjusting the display attributes, and improves the color adjustment efficiency. Since the reference image can intuitively present the color adjustment result desired by the user, adjusting the original multimedia data based on the display attributes of the reference image can improve the adjustment accuracy and avoid repeated modification of parameters.
图1为本公开实施例所提供的一种多媒体数据的处理方法的流程示意图,本公开实施例适用于视频调色的情形,该方法可以由多媒体数据的处理装置来执行,该装置可以通过软件和/或硬件的形式实现,可选的,通过电子设备来实现,该电子设备可以是移动终端、PC端或服务器等。Figure 1 is a flow chart of a method for processing multimedia data provided by an embodiment of the present disclosure. The embodiment of the present disclosure is applicable to the case of video color grading. The method can be executed by a multimedia data processing device, which can be implemented in the form of software and/or hardware. Optionally, it can be implemented by an electronic device, which can be a mobile terminal, a PC or a server, etc.
如图1所示,所述方法包括:As shown in FIG1 , the method comprises:
S110、获取原始多媒体数据。原始多媒体数据可以为用户选定的待进行调色的视频或图像等。一些实施例中,可以指定客户端中的历史视频作为原始多媒体数据。或者,可以获取用户实时拍摄的多媒体数据作为原始多媒体数据。或者,还可以获取用户从本地相册选择的多媒体数据作为原始多媒体数据。示例性地,获取原始多媒体数据,并播放原始多媒体数据。在不中断播放的情况下调整原始多媒体数据的显示属性。S110, obtaining original multimedia data. The original multimedia data may be a video or image selected by the user to be color-adjusted, etc. In some embodiments, a historical video in the client may be specified as the original multimedia data. Alternatively, multimedia data captured by the user in real time may be obtained as the original multimedia data. Alternatively, multimedia data selected by the user from a local album may be obtained as the original multimedia data. Exemplarily, the original multimedia data is obtained and played. The display properties of the original multimedia data are adjusted without interrupting the playback.
S120、获取参考图像。参考图像表征调整原始多媒体数据的图像帧的显示属性的参照图像。本公开实施例中,参考图像可以为用户上传的满足调色要求的图像。所述参考图像可以为用户拍摄的图像,也可以为从互联网下载的图像,本公开实施例并不限定参考图像的来源。示例性地,检测到调色功能开启,并获取到用户选定的待调色的原始多媒体数据之后,获取用户上传的参考图像。由于参考图像的显示风格满足用户的调色需求,可以避免手动调色过程中,为了得到满足调色需求的调色结果而反复修改参数的步骤,提升的调色效率。S120, obtain a reference image. The reference image represents a reference image for adjusting the display properties of the image frame of the original multimedia data. In the embodiment of the present disclosure, the reference image may be an image uploaded by a user that meets the color adjustment requirements. The reference image may be an image taken by a user, or may be an image downloaded from the Internet, and the embodiment of the present disclosure does not limit the source of the reference image. Exemplarily, after detecting that the color adjustment function is turned on and obtaining the original multimedia data to be color adjusted selected by the user, the reference image uploaded by the user is obtained. Since the display style of the reference image meets the color adjustment requirements of the user, the step of repeatedly modifying parameters in order to obtain a color adjustment result that meets the color adjustment requirements during the manual color adjustment process can be avoided, thereby improving the color adjustment efficiency.
S130、根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据,显示所述目标多媒体数据。所述显示属性表征所述参考图像在色彩和亮度等方面的显示效果。例如,显示属性可以包括亮度属性和色彩属性等。目标多媒体数据可以 表征原始多媒体数据的调色结果。所述目标多媒体数据包括所述原始多媒体数据中基于所述参考图像的显示属性调整后的图像帧序列,所述目标多媒体数据呈现与参考图像相似的显示效果。示例性地,对于所述参考图像和原始多媒体数据中的图像帧,分别根据所述显示属性确定图像对应的目标像素区间。S130, adjusting the image frame in the original multimedia data according to the display attribute of the reference image to obtain target multimedia data, and displaying the target multimedia data. The display attribute represents the display effect of the reference image in terms of color and brightness. For example, the display attribute may include brightness attribute and color attribute. The target multimedia data may be Characterize the color adjustment result of the original multimedia data. The target multimedia data includes a sequence of image frames in the original multimedia data adjusted based on the display attributes of the reference image, and the target multimedia data presents a display effect similar to the reference image. Exemplarily, for the image frames in the reference image and the original multimedia data, the target pixel interval corresponding to the image is determined according to the display attributes respectively.
对于每个所述目标像素区间,基于所述参考图像的像素的显示属性对所述原始多媒体数据中的图像帧进行属性匹配。融合每个所述目标像素区间对应的属性匹配结果,得到所述图像帧对应的多媒体数据处理结果。属性匹配可以包括对色彩和亮度等属性进行匹配。属性匹配结果表征将参考图像的属性特征迁移到原始多媒体数据的图像帧的处理结果。通过将参考图像的属性特征迁移到原始多媒体数据的图像帧,使得原始多媒体数据的图像帧呈现与参考图像相似的显示属性。对于所述原始多媒体数据的图像帧中肤色区域中的每个像素,根据所述参考图像的显示属性调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果。对于所述原始多媒体数据中的图像帧,根据肤色权重融合所述肤色调整结果和多媒体数据处理结果,得到目标多媒体数据。所述肤色权重表征原始多媒体数据的图像帧中每个像素属于肤色区域的概率。For each target pixel interval, attribute matching is performed on the image frame in the original multimedia data based on the display attributes of the pixels of the reference image. The attribute matching results corresponding to each target pixel interval are fused to obtain the multimedia data processing results corresponding to the image frame. Attribute matching may include matching attributes such as color and brightness. The attribute matching result represents the processing result of migrating the attribute features of the reference image to the image frame of the original multimedia data. By migrating the attribute features of the reference image to the image frame of the original multimedia data, the image frame of the original multimedia data presents display attributes similar to the reference image. For each pixel in the skin color area of the image frame of the original multimedia data, the display attributes of the pixel are adjusted according to the display attributes of the reference image to obtain the skin color adjustment result corresponding to the image frame. For the image frame in the original multimedia data, the skin color adjustment result and the multimedia data processing result are fused according to the skin color weight to obtain the target multimedia data. The skin color weight represents the probability that each pixel in the image frame of the original multimedia data belongs to the skin color area.
本公开实施例中,分别对参考图像和原始多媒体数据的图像帧进行颜色空间转换。基于颜色空间转换结果分别得到参考图像和原始多媒体数据的图像帧的亮度通道数据。根据所述亮度通道数据分别确定参考图像和原始多媒体数据的图像帧对应的直方图。根据参考图像的直方图中每个亮度值对应的像素数量,确定参考图像对应的亮度分割点。然后,根据参考图像对应的亮度分割点将所述参考图像分割为多个目标像素区间。根据原始多媒体数据的图像帧的直方图中每个亮度值对应的像素数量,确定原始多媒体数据中图像帧对应的亮度分割点。然后,根据所述图像帧对应的亮度分割点将所述图像帧分割为多个目标像素区间。例如,目标像素区间包括阴影区间、中间调区间和高光区间等。In the disclosed embodiment, the reference image and the image frame of the original multimedia data are respectively converted into color space. Based on the color space conversion result, the brightness channel data of the reference image and the image frame of the original multimedia data are respectively obtained. According to the brightness channel data, the histograms corresponding to the reference image and the image frame of the original multimedia data are respectively determined. According to the number of pixels corresponding to each brightness value in the histogram of the reference image, the brightness segmentation point corresponding to the reference image is determined. Then, according to the brightness segmentation point corresponding to the reference image, the reference image is segmented into a plurality of target pixel intervals. According to the number of pixels corresponding to each brightness value in the histogram of the image frame of the original multimedia data, the brightness segmentation point corresponding to the image frame in the original multimedia data is determined. Then, according to the brightness segmentation point corresponding to the image frame, the image frame is segmented into a plurality of target pixel intervals. For example, the target pixel intervals include shadow intervals, midtone intervals, and highlight intervals.
一些实施例中,按照亮度属性对所述参考图像的像素进行统计,根据像素统计结果确定亮度分割点。然后,根据亮度分割点将参考图像分割为若干个目标像素区间。目标像素区间的数量基于亮度分割点的数量确定,且相邻的目标像素区间具有重叠区域。In some embodiments, the pixels of the reference image are counted according to the brightness attribute, and the brightness segmentation points are determined according to the pixel statistics. Then, the reference image is segmented into a plurality of target pixel intervals according to the brightness segmentation points. The number of target pixel intervals is determined based on the number of brightness segmentation points, and adjacent target pixel intervals have overlapping areas.
一些实施例中,将参考图像转换为YUV颜色空间。确定Y通道数据为亮度属性,通过Y通道数据生成参考图像的Y通道的直方图。根据Y通道的直方图中的每个亮度值确定阴影、中间调和高光的像素。确定备选分割点以使阴影、中间调和高光对应的像素数量相同。或者,确定备选分割点以使阴影区间、中间调区间和高光区间的面积均相同。In some embodiments, the reference image is converted to a YUV color space. The Y channel data is determined to be a brightness attribute, and a histogram of the Y channel of the reference image is generated through the Y channel data. The pixels of the shadow, midtone, and highlight are determined according to each brightness value in the histogram of the Y channel. The alternative segmentation point is determined so that the number of pixels corresponding to the shadow, midtone, and highlight is the same. Alternatively, the alternative segmentation point is determined so that the areas of the shadow interval, the midtone interval, and the highlight interval are all the same.
为了避免在备份分割点对应的像素出现颜色跳变,可以在相邻区间之间错开一定范围,使相邻区间存在重叠区域。重叠区域的面积可以根据实际应用场景设定。例如,可以将重叠区域设置为参考图像对应的颜色区域的面积的n倍,且0<n<1。In order to avoid color jumps in pixels corresponding to backup segmentation points, adjacent intervals can be staggered by a certain range so that adjacent intervals have overlapping areas. The area of the overlapping area can be set according to the actual application scenario. For example, the overlapping area can be set to n times the area of the color area corresponding to the reference image, and 0<n<1.
根据每个重叠区域的起点对应的亮度值和终点对应的亮度值,确定亮度分割点。例如, 目标像素区间包括阴影区间、中间调区间和高光区间为例,则目标像素区间中存在两个重叠区域,即阴影区间与中间调区间的重叠区域和中间调区间和高光区间的重叠区域。目标像素区间中存在四个亮度分割点,分别为S1,S2,S3和S4,且S1<S2<S3<S4。阴影区间的范围为[0,S2],中间调区间的范围为[S1,S4],高光区间的范围为[S3,255]。The brightness segmentation point is determined based on the brightness value corresponding to the starting point and the brightness value corresponding to the end point of each overlapping area. For example, if the target pixel interval includes the shadow interval, the midtone interval and the highlight interval, there are two overlapping areas in the target pixel interval, namely the overlapping area between the shadow interval and the midtone interval and the overlapping area between the midtone interval and the highlight interval. There are four brightness segmentation points in the target pixel interval, namely S1, S2, S3 and S4, and S1<S2<S3<S4. The range of the shadow interval is [0, S2], the range of the midtone interval is [S1, S4], and the range of the highlight interval is [S3, 255].
采用与上述方式相似的方式确定原始多媒体数据中的图像帧的亮度分割点和目标像素区间的范围,此处不再赘述。The brightness segmentation points and the range of the target pixel interval of the image frame in the original multimedia data are determined in a manner similar to the above manner, which will not be described in detail here.
对于原始多媒体数据中的图像帧的每一个像素,若所述像素属于目标像素区间的重叠区域,则根据所述像素的亮度属性和所述重叠区域对应的亮度分割点。随后确定所述像素与每个目标像素区间对应的第一权重。或者,若所述像素属于所述目标像素区间的非重叠区域,则根据所述像素所属的目标像素区间确定所述像素与每个目标像素区间对应的第一权重。For each pixel of the image frame in the original multimedia data, if the pixel belongs to the overlapping area of the target pixel interval, then the first weight corresponding to the pixel and each target pixel interval is determined according to the brightness attribute of the pixel and the brightness segmentation point corresponding to the overlapping area. Alternatively, if the pixel belongs to the non-overlapping area of the target pixel interval, then the first weight corresponding to the pixel and each target pixel interval is determined according to the target pixel interval to which the pixel belongs.
在确定亮度分割点之后,对于原始多媒体数据中的图像帧的每个像素,根据其亮度可以确定若干个第一权重。第一权重的数量与目标像素区间的数量一致。例如,若将原始多媒体数据中的图像帧分为阴影区间、中间调区间和高光区间,可以计算原始多媒体数据中的图像帧中每个像素的三个第一权重。即分别为像素对应阴影区间的阴影权重Wshadow,像素对应中间调区间的中间调权重Wmid,以及像素对应高光区间的高光权重Whighlight,且三个权重的和为1。After determining the brightness segmentation point, for each pixel of the image frame in the original multimedia data, a number of first weights can be determined according to its brightness. The number of first weights is consistent with the number of target pixel intervals. For example, if the image frame in the original multimedia data is divided into a shadow interval, a midtone interval, and a highlight interval, three first weights of each pixel in the image frame in the original multimedia data can be calculated. That is, the shadow weight W shadow of the pixel corresponding to the shadow interval, the midtone weight W mid of the pixel corresponding to the midtone interval, and the highlight weight W highlight of the pixel corresponding to the highlight interval, and the sum of the three weights is 1.
对于原始多媒体数据中的图像帧的每个像素,若所述像素的亮度完全处于阴影区间,则该像素对应阴影区间的第一权重为Wshadow=1,该像素对应中间调区间的第一权重为Wmid=0,该像素对应高光区间的第一权重为Whighlight=0。可以采用相同的方式确定完全处于中间调区间和高光区间的像素的第一权重,此处不再赘述。For each pixel of the image frame in the original multimedia data, if the brightness of the pixel is completely in the shadow range, the first weight of the pixel corresponding to the shadow range is W shadow = 1, the first weight of the pixel corresponding to the midtone range is W mid = 0, and the first weight of the pixel corresponding to the highlight range is W highlight = 0. The first weights of pixels completely in the midtone range and the highlight range can be determined in the same manner, which will not be repeated here.
若所述像素的亮度属于阴影区间和中间调区间的重叠区域,则所述像素既有阴影权重,也有中间调权重。根据像素的亮度以及所述重叠区域的亮度分割点(S1和S2)的亮度,确定所述像素对应于阴影区间的第一权重Wshadow,确定所述像素对应于中间调区间的第一权重Wmid,确定所述像素对应于高光区间的第一权重为Whighlight=0。If the brightness of the pixel belongs to the overlapping area of the shadow interval and the midtone interval, the pixel has both a shadow weight and a midtone weight. According to the brightness of the pixel and the brightness of the brightness segmentation points (S1 and S2) of the overlapping area, the first weight W shadow of the pixel corresponding to the shadow interval is determined, the first weight W mid of the pixel corresponding to the midtone interval is determined, and the first weight W highlight of the pixel corresponding to the highlight interval is determined to be 0.
采用上述相似的方法可以确定参考图像对应的第一权重,此处不再赘述。The first weight corresponding to the reference image can be determined by using a method similar to the above, which will not be described in detail here.
一些实施例中,对于每个所述目标像素区间,基于所述参考图像的像素的显示属性对所述原始多媒体数据中的图像帧进行属性匹配。融合每个所述目标像素区间对应的属性匹配结果,得到所述图像帧对应的多媒体数据处理结果,包括:基于所述参考图像的亮度属性对所述原始多媒体数据中的图像帧进行亮度匹配,得到所述原始多媒体数据中的图像帧的亮度匹配结果。为了将图像亮度和色彩分开进行匹配,将原始多媒体数据中的图像帧转换为YUV颜色空间。对UV通道进行色彩匹配,对Y通道进行亮度匹配。例如,采用直方图匹配算法,将原始多媒体数据中的图像帧的Y通道与参考图像的Y通道进行直方图匹 配,得到亮度匹配结果。In some embodiments, for each of the target pixel intervals, attribute matching is performed on the image frames in the original multimedia data based on the display attributes of the pixels of the reference image. The attribute matching results corresponding to each of the target pixel intervals are fused to obtain the multimedia data processing results corresponding to the image frames, including: performing brightness matching on the image frames in the original multimedia data based on the brightness attribute of the reference image to obtain brightness matching results of the image frames in the original multimedia data. In order to match the image brightness and color separately, the image frames in the original multimedia data are converted to the YUV color space. Color matching is performed on the UV channel, and brightness matching is performed on the Y channel. For example, a histogram matching algorithm is used to perform histogram matching on the Y channel of the image frame in the original multimedia data and the Y channel of the reference image. Match and get the brightness matching result.
根据每个所述目标像素区间的色彩属性确定所述目标像素区间对应的色彩特征和变换矩阵。所述色彩特征表征所述原始多媒体数据中的图像帧和参考图像分别对应的所述目标像素区间的像素的色彩分布。本公开实施例,色彩特征可以包括UV通道均值和协方差矩阵等。The color features and transformation matrix corresponding to the target pixel interval are determined according to the color attributes of each target pixel interval. The color features represent the color distribution of pixels in the target pixel interval corresponding to the image frame and the reference image in the original multimedia data. In the embodiment of the present disclosure, the color features may include UV channel mean and covariance matrix, etc.
对于原始多媒体数据中的图像帧和参考图像,分别获取阴影区间、中间调区间和高光区间中像素的UV通道数据。根据UV通道数据计算UV均值和协方差矩阵。然后,根据UV均值和协方差矩阵确定变换矩阵M。For the image frame and the reference image in the original multimedia data, the UV channel data of the pixels in the shadow interval, the midtone interval and the highlight interval are obtained respectively. The UV mean and the covariance matrix are calculated according to the UV channel data. Then, the transformation matrix M is determined according to the UV mean and the covariance matrix.
一些实施例中,根据每个所述目标像素区间的色彩属性确定所述目标像素区间对应的色彩特征和变换矩阵,包括:根据所述原始多媒体数据中的图像帧的每个所述目标像素区间的像素的色彩属性,确定所述目标像素区间对应的第一色彩均值和第一协方差矩阵。根据所述参考图像的每个所述目标像素区间的像素的色彩属性,确定所述目标像素区间对应的第二色彩均值和第二协方差矩阵;根据所述第一色彩均值、第二色彩均值、第一协方差矩阵和第二协方差矩阵确定所述目标像素区间对应的变换矩阵,所述色彩特征包括所述第一色彩均值和第二色彩均值。In some embodiments, the color features and transformation matrix corresponding to the target pixel interval are determined according to the color attributes of each target pixel interval, including: determining the first color mean and the first covariance matrix corresponding to the target pixel interval according to the color attributes of the pixels in each target pixel interval of the image frame in the original multimedia data. Determining the second color mean and the second covariance matrix corresponding to the target pixel interval according to the color attributes of the pixels in each target pixel interval of the reference image; determining the transformation matrix corresponding to the target pixel interval according to the first color mean, the second color mean, the first covariance matrix and the second covariance matrix, wherein the color features include the first color mean and the second color mean.
以阴影区间为例,对于所有亮度处于在阴影区间的像素,计算原始多媒体数据中的图像帧所有阴影像素的UV通道均值协方差矩阵Σsrc。第一色彩均值可以表示为和第一协方差矩阵可以表示为Σsrc。计算参考图像所有阴影像素的UV通道均值 协方差矩阵Σref。第二色彩均值可以表示为和第二协方差矩阵可以表示为Σref。采用MKL算法基于上述均值和协方差矩阵,确定变换矩阵M。Taking the shadow interval as an example, for all pixels whose brightness is in the shadow interval, the UV channel mean of all shadow pixels in the image frame in the original multimedia data is calculated. Covariance matrix Σ src . The first color mean can be expressed as and The first covariance matrix can be expressed as Σ src . Calculate the UV channel mean of all shadow pixels in the reference image The covariance matrix Σ ref . The second color mean can be expressed as and The second covariance matrix can be expressed as Σ ref . The transformation matrix M is determined based on the above mean and covariance matrix using the MKL algorithm.
一些实施例中,在确定所述目标像素区间对应的第一色彩均值和第一协方差矩阵之后,还包括:确定所述第一协方差矩阵的特征值。所述特征值表征所述原始多媒体数据中的图像帧的像素按照高斯分布呈现的椭圆的形状。在所述特征值满足设定条件的情况下,将单位矩阵作为所述变换矩阵。In some embodiments, after determining the first color mean and the first covariance matrix corresponding to the target pixel interval, the method further includes: determining an eigenvalue of the first covariance matrix. The eigenvalue represents the shape of an ellipse presented by the pixels of the image frame in the original multimedia data according to the Gaussian distribution. When the eigenvalue satisfies the set conditions, the unit matrix is used as the transformation matrix.
预设条件表征原始多媒体数据中的图像帧的像素按照高斯分布呈现的椭圆的形状阈值,对于超过形状阈值的高斯分布进行色彩变换,则可能产生色彩异常点。在这种情况下,放弃基于协方差矩阵计算变换矩阵,而是把变换矩阵M设置为单位矩阵。The preset condition represents the shape threshold of the ellipse presented by the Gaussian distribution of the pixels of the image frame in the original multimedia data. If the Gaussian distribution exceeding the shape threshold is subjected to color transformation, color abnormal points may be generated. In this case, the transformation matrix is calculated based on the covariance matrix, and the transformation matrix M is set to the unit matrix instead.
例如,根据原始多媒体数据中的图像帧的协方差矩阵Σsrc确定协方差矩阵的特征值e1和e2,且e1>e2。协方差矩阵的特征值表示了原始多媒体数据中的图像帧的像素按照高斯分布呈现的椭圆形的长轴e1和短轴e2。如果e1/e2>设定形状阈值,则说明高斯分布呈现的椭圆形很扁。此时如果对该高斯分布进行色彩变换,则容易产生色彩异常点。因此将变换矩阵M设为单位矩阵。 For example, the eigenvalues e1 and e2 of the covariance matrix are determined according to the covariance matrix Σsrc of the image frame in the original multimedia data, and e1>e2. The eigenvalues of the covariance matrix represent the major axis e1 and minor axis e2 of the ellipse presented by the pixels of the image frame in the original multimedia data according to the Gaussian distribution. If e1/e2> the set shape threshold, it means that the ellipse presented by the Gaussian distribution is very flat. At this time, if the Gaussian distribution is subjected to color transformation, color outliers are likely to be generated. Therefore, the transformation matrix M is set to the unit matrix.
采用上述方式,可以分别计算出原始多媒体数据中的图像帧在阴影区间、中间调区间和高光区间的第一色彩均值,即为和和和分别计算出原始多媒体数据中的图像帧在阴影区间、中间调区间和高光区间的协方差矩阵,即为Σsrc0、Σsrc1,Σsrc2。分别计算出原始多媒体数据中的图像帧在阴影区间、中间调区间和高光区间的变换矩阵M0,M1,和M2。By adopting the above method, the first color mean values of the image frames in the original multimedia data in the shadow interval, the mid-tone interval and the highlight interval can be calculated respectively, which are: and and and The covariance matrices of the image frames in the original multimedia data in the shadow interval, the midtone interval and the highlight interval are calculated respectively, namely Σ src0 , Σ src1 , Σ src2 . The transformation matrices M 0 , M 1 , and M 2 of the image frames in the original multimedia data in the shadow interval, the midtone interval and the highlight interval are calculated respectively.
根据所述色彩特征和变换矩阵分别对所述原始多媒体数据中的图像帧对应的每个所述目标像素区间进行色彩匹配。Color matching is performed on each of the target pixel intervals corresponding to the image frame in the original multimedia data according to the color features and the transformation matrix.
示例性地,对于所述原始多媒体数据中的图像帧的每个目标像素区间,根据所述目标像素区间中像素的色彩信息和第一色彩均值确定所述像素对应的第二权重。Exemplarily, for each target pixel interval of the image frame in the original multimedia data, the second weight corresponding to the pixel is determined according to color information of the pixel in the target pixel interval and the first color mean value.
第二权重表征像素的色彩与高斯分布中的均值的偏离程度。第二权重越小,表示像素的色彩值偏离高斯分布的均值。第二权重越大,表示像素的色彩值靠近高斯分布的均值。The second weight represents the degree of deviation of the color of the pixel from the mean in the Gaussian distribution. The smaller the second weight, the more the color value of the pixel deviates from the mean of the Gaussian distribution. The larger the second weight, the closer the color value of the pixel is to the mean of the Gaussian distribution.
以阴影区间为例。原始多媒体数据中的图像帧的阴影区间的像素在UV通道的色彩信息uv可以表示为二维向量,原始多媒体数据中的图像帧的阴影区间的第一色彩均值可以表示为根据像素的色彩信息uv、第一色彩均值和第一协方差矩阵Σsrc0计算第二权重。采用相似的方式,计算中间调区间和高光区间中像素对应的第二权重,此处不再赘述。Take the shadow interval as an example. The color information uv of the pixels in the shadow interval of the image frame in the original multimedia data in the UV channel can be expressed as a two-dimensional vector, The first color mean value of the shadow interval of the image frame in the original multimedia data It can be expressed as According to the pixel color information uv, the first color mean The second weight is calculated by using the first covariance matrix Σ src0 . In a similar manner, the second weights corresponding to the pixels in the midtone interval and the highlight interval are calculated, which will not be repeated here.
根据所述色彩特征和变换矩阵分别对所述目标像素区间进行色彩匹配,得到所述目标像素区间对应的第一色彩匹配结果和第二色彩匹配结果。例如,可以基于变换矩阵对原始多媒体数据中的图像帧的目标像素区间进行色彩匹配,得到第一色彩匹配结果。第一色彩匹配结果可以为基于变换矩阵对原始多媒体数据中的图像帧的目标像素区间进行色彩变换后,得到的新的色彩信息。基于原始多媒体数据中的图像帧中所述像素的色彩、所述像素所属的目标像素区间对应的第一色彩均值,参考图像中所述目标像素区间对应的第二色彩均值,确定第二色彩匹配结果。第二色彩匹配结果可以为基于原始多媒体数据中的图像帧和参考图像在目标像素区间的色彩均值,调整所述目标像素区间的色彩信息,得到的新的色彩信息。Color matching is performed on the target pixel interval according to the color feature and the transformation matrix, respectively, to obtain a first color matching result and a second color matching result corresponding to the target pixel interval. For example, color matching can be performed on the target pixel interval of the image frame in the original multimedia data based on the transformation matrix to obtain a first color matching result. The first color matching result can be new color information obtained after color transformation is performed on the target pixel interval of the image frame in the original multimedia data based on the transformation matrix. The second color matching result is determined based on the color of the pixel in the image frame in the original multimedia data, the first color mean corresponding to the target pixel interval to which the pixel belongs, and the second color mean corresponding to the target pixel interval in the reference image. The second color matching result can be new color information obtained by adjusting the color information of the target pixel interval based on the color mean of the image frame in the original multimedia data and the reference image in the target pixel interval.
以阴影区间为例。基于变换矩阵对原始多媒体数据中的图像帧在阴影区间中每个像素进行色彩变换,得到新的色彩信息,作为第一色彩匹配结果。基于原始多媒体数据中的图像帧和参考图像在阴影区间的色彩均值调整原始多媒体数据中的图像帧在阴影区间中每个像素的色彩信息,得到新的色彩信息,作为第二色彩匹配结果。例如,原始多媒体数据中的图像帧的像素的色彩信息uv结合阴影区间对应的变换矩阵M0,得到阴影区间对应的第一色彩匹配结果,实现将参考图像的色彩风格迁移至原始多媒体数据中的图像帧。原始多媒体数据中的图像帧的像素的色彩信息uv减去原始多媒体数据中的图像帧的后叠 加上参考图像的得到阴影区间对应的第二色彩匹配结果。表示参考图像的阴影区间的第二色彩均值,可以表示为 Take the shadow interval as an example. Perform a color transformation on each pixel in the shadow interval of the image frame in the original multimedia data based on the transformation matrix to obtain new color information as the first color matching result. Adjust the color information of each pixel in the shadow interval of the image frame in the original multimedia data based on the color mean of the image frame in the original multimedia data and the reference image in the shadow interval to obtain new color information as the second color matching result. For example, the color information uv of the pixel of the image frame in the original multimedia data is combined with the transformation matrix M 0 corresponding to the shadow interval to obtain the first color matching result corresponding to the shadow interval, thereby realizing the color style transfer of the reference image to the image frame in the original multimedia data. The color information uv of the pixel of the image frame in the original multimedia data minus the color information uv of the pixel of the image frame in the original multimedia data Back Fold With reference image A second color matching result corresponding to the shadow interval is obtained. The second color mean of the shadow interval of the reference image can be expressed as
根据所述第二权重融合所述第一色彩匹配结果和第二色彩匹配结果,得到所述目标像素区间的色彩匹配结果。The first color matching result and the second color matching result are fused according to the second weight to obtain a color matching result of the target pixel interval.
以阴影区间为例。根据第二权重确定融合系数,根据融合系数融合阴影区间对应的第一色彩匹配结果和第二色彩匹配结果,得到阴影区间的色彩匹配结果。融合系数包括第二权重,且各融合系数的和为1。采用相似方式计算中间调区间的色彩匹配结果和高光区间的色彩匹配结果,此处不再赘述。Take the shadow interval as an example. The fusion coefficient is determined according to the second weight, and the first color matching result and the second color matching result corresponding to the shadow interval are fused according to the fusion coefficient to obtain the color matching result of the shadow interval. The fusion coefficient includes the second weight, and the sum of the fusion coefficients is 1. The color matching results of the midtone interval and the color matching results of the highlight interval are calculated in a similar manner, which will not be repeated here.
需要说明的是,对于原始多媒体数据中的图像帧中亮度完全属于阴影区间、中间调区间或高光区间的像素,仅存在阴影区间的色彩匹配结果。对于原始多媒体数据中的图像帧中亮度属于阴影区间和中间调区间的重叠区域的像素,存在阴影区间的色彩匹配结果和中间调区间的色彩匹配结果。对于原始多媒体数据中的图像帧中亮度属于中间调区间和高光调区间的重叠区域的像素,存在中间调区间色彩匹配结果和高光区间的色彩匹配结果。It should be noted that, for pixels in the image frame in the original multimedia data whose brightness completely belongs to the shadow interval, the midtone interval or the highlight interval, there is only a color matching result for the shadow interval. For pixels in the image frame in the original multimedia data whose brightness belongs to the overlapping area of the shadow interval and the midtone interval, there are a color matching result for the shadow interval and a color matching result for the midtone interval. For pixels in the image frame in the original multimedia data whose brightness belongs to the overlapping area of the midtone interval and the highlight interval, there are a color matching result for the midtone interval and a color matching result for the highlight interval.
对于所述原始多媒体数据中的图像帧的每个像素,根据所述像素与每个目标像素区间对应的第一权重融合所述目标像素区间的色彩匹配结果,得到所述原始多媒体数据中的图像帧的色彩匹配结果;根据所述原始多媒体数据中的图像帧的色彩匹配结果和亮度匹配结果确定所述原始多媒体数据中的图像帧的处理结果图像,作为图像帧对应的多媒体数据处理结果。For each pixel of the image frame in the original multimedia data, the color matching result of the target pixel interval is fused according to the first weight corresponding to the pixel and each target pixel interval to obtain the color matching result of the image frame in the original multimedia data; and the processing result image of the image frame in the original multimedia data is determined according to the color matching result and brightness matching result of the image frame in the original multimedia data as the multimedia data processing result corresponding to the image frame.
一些实施例中,对于原始多媒体数据中的图像帧的每个像素,首先确定该像素分别与各个目标像素区间(即阴影区间、中间调区间和高光区间)对应的第一权重(即阴影权重、中间调权重和高光权重)。根据第一权重,融合该像素在对应的目标像素区间(即阴影权重、中间调权重和高光权重)的色彩匹配结果,得到原始多媒体数据中的图像帧的色彩匹配结果。In some embodiments, for each pixel of the image frame in the original multimedia data, firstly, the first weights (i.e., shadow weight, midtone weight, and highlight weight) corresponding to the pixel and the respective target pixel intervals (i.e., shadow interval, midtone interval, and highlight interval) are determined. According to the first weights, the color matching results of the pixel in the corresponding target pixel intervals (i.e., shadow weight, midtone weight, and highlight weight) are fused to obtain the color matching result of the image frame in the original multimedia data.
例如,对于原始多媒体数据中的图像帧的每个像素,根据所述像素的阴影权重、中间调权重和高光权重,融合该像素在阴影区间的色彩匹配结果、该像素在中间调区间的色彩匹配结果以及该像素在高光调区间的色彩匹配结果。由此,得到原始多媒体数据中的图像帧的色彩匹配结果。For example, for each pixel of the image frame in the original multimedia data, according to the shadow weight, midtone weight and highlight weight of the pixel, the color matching result of the pixel in the shadow interval, the color matching result of the pixel in the midtone interval and the color matching result of the pixel in the highlight interval are fused. Thus, the color matching result of the image frame in the original multimedia data is obtained.
一些实施例中,对于所述原始多媒体数据的图像帧中肤色区域中的像素,根据所述参考图像的肤色区域的显示属性调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果。包括:确定所述原始多媒体数据的图像帧中每个像素属于肤色区域的概率,作为肤色权重,根据所述肤色权重确定所述原始多媒体数据的图像帧中肤色区域包含的像素。根据所述原始多媒体数据的图像帧中肤色区域包含的像素的显示属性,确定所述原始多媒体 数据的图像帧中肤色区域的显示属性均值。对于所述原始多媒体数据的图像帧中肤色区域中的像素,根据所述图像帧和参考图像分别在肤色区域的显示属性均值调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果。In some embodiments, for pixels in a skin color area in an image frame of the original multimedia data, the display attributes of the pixels are adjusted according to the display attributes of the skin color area of the reference image to obtain a skin color adjustment result corresponding to the image frame. This includes: determining the probability that each pixel in the image frame of the original multimedia data belongs to a skin color area as a skin color weight, and determining the pixels included in the skin color area in the image frame of the original multimedia data according to the skin color weight. Determining the original multimedia data according to the display attributes of the pixels included in the skin color area in the image frame of the original multimedia data For pixels in the skin color area in the image frame of the original multimedia data, the display attributes of the pixels are adjusted according to the mean values of the display attributes of the image frame and the reference image in the skin color area, respectively, to obtain a skin color adjustment result corresponding to the image frame.
肤色调整结果表征将参考图像的色彩和亮度特征迁移到原始多媒体数据的图像帧中肤色区域的处理结果。The skin color adjustment result represents the processing result of migrating the color and brightness characteristics of the reference image to the skin color area in the image frame of the original multimedia data.
示例性地,将原始多媒体数据输入肤色分割模型。采用肤色分割模型分别对原始多媒体数据中的图像帧进行肤色分割,得到所述图像帧中每个像素属于肤色区域的概率。根据所述概率确定原始多媒体数据中的图像帧的肤色区域包含的像素。相似地,将参考图像输入肤色分割模型。采用肤色分割模型对所述参考图像进行肤色分割,得到参考图像中每个像素属于肤色区域的概率。根据所述概率确定所述参考图像的肤色区域包含的像素。Exemplarily, the original multimedia data is input into the skin color segmentation model. The skin color segmentation model is used to perform skin color segmentation on the image frames in the original multimedia data, and the probability that each pixel in the image frame belongs to the skin color area is obtained. The pixels contained in the skin color area of the image frame in the original multimedia data are determined according to the probability. Similarly, the reference image is input into the skin color segmentation model. The skin color segmentation model is used to perform skin color segmentation on the reference image, and the probability that each pixel in the reference image belongs to the skin color area is obtained. The pixels contained in the skin color area of the reference image are determined according to the probability.
根据原始多媒体数据中的图像帧的肤色区域包含的每个像素的亮度,确定所述原始多媒体数据中的图像帧的肤色区域的亮度均值。根据参考图像的肤色区域包含的每个像素的亮度,确定所述参考图像中肤色区域的亮度均值。根据原始多媒体数据中的图像帧的肤色区域包含的每个像素的色度,确定所述原始多媒体数据中的图像帧的肤色区域的色度均值。根据参考图像的肤色区域包含的每个像素的色度,确定所述参考图像中肤色区域的色度均值。According to the brightness of each pixel contained in the skin color area of the image frame in the original multimedia data, the brightness mean of the skin color area of the image frame in the original multimedia data is determined. According to the brightness of each pixel contained in the skin color area of the reference image, the brightness mean of the skin color area of the reference image is determined. According to the chromaticity of each pixel contained in the skin color area of the image frame in the original multimedia data, the chromaticity mean of the skin color area of the image frame in the original multimedia data is determined. According to the chromaticity of each pixel contained in the skin color area of the reference image, the chromaticity mean of the skin color area of the reference image is determined.
对于所述原始多媒体数据的图像帧中肤色区域中的每个像素,计算所述像素的亮度与所述肤色区域的亮度均值的差值。将所述差值与所述参考图像的肤色区域的亮度均值进行叠加,得到所述像素的亮度调整结果。For each pixel in the skin color area in the image frame of the original multimedia data, the difference between the brightness of the pixel and the average brightness of the skin color area is calculated, and the difference is superimposed with the average brightness of the skin color area of the reference image to obtain a brightness adjustment result of the pixel.
对于所述原始多媒体数据的图像帧中肤色区域中的每个像素,计算所述像素的色度与所述肤色区域的色度均值的差值。将所述差值与所述参考图像的肤色区域的色度均值进行叠加,得到所述像素的色度调整结果。For each pixel in the skin color area of the image frame of the original multimedia data, the difference between the chromaticity of the pixel and the chromaticity mean of the skin color area is calculated, and the difference is superimposed with the chromaticity mean of the skin color area of the reference image to obtain the chromaticity adjustment result of the pixel.
根据所述原始多媒体数据的图像帧中肤色区域中的每个像素的亮度调整结果和色度调整结果,得到所述图像帧对应的肤色调整结果。According to the brightness adjustment result and the chromaticity adjustment result of each pixel in the skin color area in the image frame of the original multimedia data, a skin color adjustment result corresponding to the image frame is obtained.
基于肤色权重采用线性融合方式对所述图像帧的多媒体数据处理结果和肤色调整结果进行融合,得到目标多媒体数据。例如,根据目标多媒体数据包括的图像帧序列对应的多媒体数据处理结果和肤色调整结果的融合结果,确定目标多媒体数据。The target multimedia data is obtained by fusing the multimedia data processing result and the skin color adjustment result of the image frame in a linear fusion manner based on the skin color weight. For example, the target multimedia data is determined based on the fusion result of the multimedia data processing result and the skin color adjustment result corresponding to the image frame sequence included in the target multimedia data.
本公开实施例的技术方案,通过参考图像的显示属性调整原始多媒体数据中图像帧,得到目标多媒体数据并显示。由此,实现了基于参考图像的色彩风格调整原始多媒体数据的显示风格,使目标多媒体数据呈现与参考图像相似的显示效果,省略了手动调整显示属性的步骤,提高了调色效率。由于参考图像能够直观的呈现用户想要的调色结果,因此基于参考图像的显示属性调整原始多媒体数据可以提高调整准确率,避免反复修改参数。The technical solution of the disclosed embodiment adjusts the image frame in the original multimedia data by the display attribute of the reference image, obtains the target multimedia data and displays it. Thus, the display style of the original multimedia data is adjusted based on the color style of the reference image, so that the target multimedia data presents a display effect similar to that of the reference image, omitting the step of manually adjusting the display attribute, and improving the color adjustment efficiency. Since the reference image can intuitively present the color adjustment result desired by the user, adjusting the original multimedia data based on the display attribute of the reference image can improve the adjustment accuracy and avoid repeated modification of parameters.
图2为本公开实施例所提供的另一种多媒体数据的处理方法的流程示意图,本公开实 施例在上述实施例的基础上,附加参考图像和特效元素的显示步骤。如图2所示,所述方法包括:S210、获取原始多媒体数据。S220、获取参考图像。S230、根据所述参考图像生成特效元素,在所述原始多媒体数据中叠加显示所述参考图像和特效元素。FIG. 2 is a flow chart of another method for processing multimedia data provided by an embodiment of the present disclosure. Embodiment Based on the above embodiment, additional display steps of reference images and special effect elements are added. As shown in FIG2 , the method includes: S210, obtaining original multimedia data. S220, obtaining a reference image. S230, generating special effect elements according to the reference image, and superimposing and displaying the reference image and special effect elements on the original multimedia data.
若在基于参考图像调整原始多媒体数据中图像帧的显示属性的过程中只显示原始多媒体数据,显示效果比较单一,用户体验不佳。甚至可能导致用户误以为未成功上传参考图像而反复操作,导致误处理。通过在原始多媒体数据中叠加显示参考图像和特效元素可以避免显示效果单一及反复上传参考图像的误操作。If only the original multimedia data is displayed during the process of adjusting the display properties of the image frame in the original multimedia data based on the reference image, the display effect is relatively simple and the user experience is poor. It may even cause the user to mistakenly believe that the reference image has not been successfully uploaded and repeat the operation, resulting in erroneous processing. By superimposing the reference image and special effect elements on the original multimedia data, the single display effect and the erroneous operation of repeatedly uploading the reference image can be avoided.
所述特效元素叠加在所述参考图像的设定位置显示并按照设定的方向移动。例如,特效元素包括色块特效等。色块特效可以包括不同颜色的色块。色块的颜色代表通过对参考图像的像素的色彩进行聚类,得到的色彩类别。或者,色块的颜色还可以为固定颜色,本公开实施例对色块的形状和颜色并不作具体限定。The special effect element is superimposed on the set position of the reference image and is displayed and moves in a set direction. For example, the special effect element includes a color block special effect, etc. The color block special effect may include color blocks of different colors. The color of the color block represents the color category obtained by clustering the colors of the pixels of the reference image. Alternatively, the color of the color block may also be a fixed color, and the embodiment of the present disclosure does not specifically limit the shape and color of the color block.
示例性地,根据所述参考图像中像素的色彩属性确定所述参考图像包含的色彩类别。生成与所述色彩类别表征的色彩对应的特效元素。Exemplarily, the color category included in the reference image is determined according to the color attribute of the pixels in the reference image, and a special effect element corresponding to the color represented by the color category is generated.
一些实施例中,预先设置多个聚类中心,并根据像素的色彩与聚类中心对应像素的色彩对参考图像中像素进行聚类,得到多个像素组。可以将所有组像素的聚类中心对应的颜色作为参考图像包含的色彩类别。In some embodiments, multiple cluster centers are pre-set, and pixels in the reference image are clustered according to the colors of the pixels and the colors of the pixels corresponding to the cluster centers to obtain multiple pixel groups. The colors corresponding to the cluster centers of all groups of pixels can be used as the color categories included in the reference image.
示例性地,在所述原始多媒体数据的图像帧中叠加显示所述参考图像,并呈现沿设定方向滑入所述原始多媒体数据的显示画面的效果。图3为本公开实施例提供的一种原始多媒体数据的图像帧中叠加显示参考图像的示意图。如图3所示,参考图像在原始多媒体数据的播放过程中,t1时刻参考图像叠加显示在原始多媒体数据的第一位置310,并在t2时刻滑动至第二位置320。Exemplarily, the reference image is superimposed and displayed in the image frame of the original multimedia data, and an effect of sliding into the display screen of the original multimedia data along a set direction is presented. FIG3 is a schematic diagram of superimposing and displaying a reference image in an image frame of original multimedia data provided by an embodiment of the present disclosure. As shown in FIG3, during the playback of the original multimedia data, the reference image is superimposed and displayed at a first position 310 of the original multimedia data at time t1 , and slides to a second position 320 at time t2 .
在所述参考图像中叠加显示线条素材,所述线条素材在所述参考图像中沿第一方向移动。线条素材在参考图像中沿第一方向移动可以呈现出光线扫描参考图像的效果。通过将线条素材沿第一方向移动的动图叠加至原始多媒体数据中显示的参考图像,呈现出光线扫描参考图像的效果。图4为本公开实施例提供的一种参考图像叠加线条素材的示意图。如图4所示,根据参考图像的宽和高生成设定透明度的遮罩图410,且遮罩图410中存在亮区411和暗区412。根据多张亮区411位置不同的遮罩图410合成动图。将动图叠加至参考图,得到光线扫描参考图像的效果。Line material is superimposed and displayed in the reference image, and the line material moves along a first direction in the reference image. The movement of the line material in the reference image along the first direction can present an effect of light scanning the reference image. By superimposing a dynamic image of the line material moving along the first direction onto the reference image displayed in the original multimedia data, an effect of light scanning the reference image is presented. FIG4 is a schematic diagram of a reference image superimposed with line material provided by an embodiment of the present disclosure. As shown in FIG4 , a mask image 410 with set transparency is generated according to the width and height of the reference image, and there are bright areas 411 and dark areas 412 in the mask image 410. A dynamic image is synthesized based on multiple mask images 410 with different positions of bright areas 411. The dynamic image is superimposed on the reference image to obtain the effect of light scanning the reference image.
按照设定叠加形式将所述特效元素叠加到所述参考图像中的第一区域。设定叠加形式表征将色块特效叠加到参考图像的形式。例如,设定叠加形式可以是按照色块的颜色依次在参考图像的不同位置显示色块。第一区域可以为参考图像范围内的预设区域。图5为本公开实施例提供的一种色块特效的示意图。如图5所示,在参考图像510中逐个显示各个色块520,直至所有色块均520显示在参考图像510中的第一区域。在每个色块520的显 示过程中,色块的宽度和高度由第一状态变换为第二状态。The special effect element is superimposed on the first area of the reference image according to the set superposition form. The set superposition form represents the form of superimposing the color block special effect on the reference image. For example, the set superposition form can be to display the color blocks in different positions of the reference image in sequence according to the color of the color blocks. The first area can be a preset area within the reference image. FIG5 is a schematic diagram of a color block special effect provided by an embodiment of the present disclosure. As shown in FIG5, each color block 520 is displayed one by one in the reference image 510 until all the color blocks 520 are displayed in the first area of the reference image 510. During the display process, the width and height of the color block are changed from a first state to a second state.
示例性地,根据所述设定叠加形式确定所述特效元素对应的显示位置和亮度排列顺序。根据所述亮度排列顺序确定所述特效元素的显示顺序。按照所述显示位置和显示顺序,在所述原始多媒体数据中图像帧中的参考图像的第一区域中叠加显示所述特效元素。显示位置表征特效元素在参考图像中的位置。例如,显示位置可以包括色块在参考图像内横向(或纵向或倾斜)排列,且相邻色块的间距相同(或逐渐增多或逐渐减小)。亮度排列顺序可以包括色彩或亮度由深到浅排列,或者,由浅到深排列等。由于每个色块表征参考图像中出现的颜色,且色块的色彩深浅不同,亮度不同。可以根据设定叠加形式中规定的亮度排序顺序确定色块在参考图像中的显示顺序。然后,按照显示顺序在色块对应的显示位置上逐个显示色块。在每个色块显示时,先在参考图像中显示色块的第一状态,且呈现该色块由第一状态转换为第二状态的过程。例如,首先在参考图像中显示一个小色块,再对小色块进行放大,得到设定宽度和高度的色块。Exemplarily, the display position and brightness arrangement order corresponding to the special effect element are determined according to the set superposition form. The display order of the special effect element is determined according to the brightness arrangement order. According to the display position and display order, the special effect element is superimposed and displayed in the first area of the reference image in the image frame in the original multimedia data. The display position represents the position of the special effect element in the reference image. For example, the display position may include color blocks arranged horizontally (or vertically or obliquely) in the reference image, and the spacing between adjacent color blocks is the same (or gradually increases or gradually decreases). The brightness arrangement order may include color or brightness arranged from dark to light, or from light to dark, etc. Since each color block represents the color appearing in the reference image, and the color blocks have different color depths and different brightness. The display order of the color blocks in the reference image can be determined according to the brightness sorting order specified in the set superposition form. Then, the color blocks are displayed one by one at the display position corresponding to the color blocks according to the display order. When each color block is displayed, the first state of the color block is first displayed in the reference image, and the process of the color block converting from the first state to the second state is presented. For example, a small color block is first displayed in the reference image, and then the small color block is enlarged to obtain a color block with a set width and height.
将所述特效元素从所述参考图像的第一区域沿第二方向移动至所述参考图像对应的第二区域,调整所述特效元素在所述第二区域的显示状态。第二区域表征参考图像外的区域范围。图6为本公开实施例提供的一种色块移位的示意图。如图6所示,在所有色块显示在参考图像的第一区域之后,将所有色块610由横向排列旋转为纵向排列,并移动至参考图像对应的第二区域620。在所有色块610达到第二区域620之后,增大相邻色块之间的间距。S240、根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据。S250、在所述特效元素在所述原始多媒体数据中的显示状态满足设定条件时,显示所述目标多媒体数据。Move the special effect element from the first area of the reference image along the second direction to the second area corresponding to the reference image, and adjust the display state of the special effect element in the second area. The second area represents the area range outside the reference image. Figure 6 is a schematic diagram of a color block shift provided by an embodiment of the present disclosure. As shown in Figure 6, after all color blocks are displayed in the first area of the reference image, all color blocks 610 are rotated from a horizontal arrangement to a vertical arrangement, and moved to the second area 620 corresponding to the reference image. After all color blocks 610 reach the second area 620, the spacing between adjacent color blocks is increased. S240. Adjust the image frame in the original multimedia data according to the display attributes of the reference image to obtain target multimedia data. S250. When the display state of the special effect element in the original multimedia data meets the set conditions, display the target multimedia data.
设定条件可以表征特效元素从原始多媒体数据中消失的条件,在特效元素从原始多媒体数据中消失时,显示目标多媒体数据。例如,当增大相邻色块之间的间距,使相邻色块之间的间距为设定数值时,确定特效色彩满足设定条件。或者,当增大相邻色块之间的间距,并且色块的显示时间达到设定时间时,确定特效色彩满足设定条件。The set condition may represent a condition under which the special effect element disappears from the original multimedia data, and when the special effect element disappears from the original multimedia data, the target multimedia data is displayed. For example, when the spacing between adjacent color blocks is increased so that the spacing between adjacent color blocks is a set value, it is determined that the special effect color meets the set condition. Alternatively, when the spacing between adjacent color blocks is increased and the display time of the color blocks reaches a set time, it is determined that the special effect color meets the set condition.
本公开实施例的技术方案,通过生成参考图像对应的特效元素,并在原始多媒体数据中叠加显示参考图像和特效元素,可以直观地展示用户选定的参考图像的色彩风格,并且能够通过生动地展示特效元素,避免在基于参考图像调整原始多媒体数据的显示属性的过程中,存在显示效果单一或反复上传参考图像的误操作,增加互动性和趣味性,提升用户体验。The technical solution of the disclosed embodiment can intuitively display the color style of the reference image selected by the user by generating special effect elements corresponding to the reference image and superimposing the reference image and the special effect elements in the original multimedia data. It can also avoid erroneous operations such as a single display effect or repeated uploading of reference images in the process of adjusting the display properties of the original multimedia data based on the reference image by vividly displaying the special effect elements, thereby increasing interactivity and fun and improving user experience.
图7为本公开实施例所提供的一种多媒体数据的处理装置结构示意图,该装置可以通过软件和/或硬件的形式实现,可选的,通过电子设备来实现,该电子设备可以是移动终端、PC端或服务器等。7 is a schematic diagram of the structure of a multimedia data processing device provided in an embodiment of the present disclosure. The device may be implemented in the form of software and/or hardware. Optionally, it may be implemented by an electronic device, which may be a mobile terminal, a PC, a server, etc.
如图7所示,所述装置包括:图像获取模块710以及图像调整模块720。数据获取模 块710,用于获取原始多媒体数据;图像获取模块720,用于获取参考图像;图像调整模块730,用于根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据,显示所述目标多媒体数据。所述目标多媒体数据包括所述原始多媒体数据中基于所述参考图像的显示属性调整后的图像帧序列。As shown in FIG7 , the device includes: an image acquisition module 710 and an image adjustment module 720. Block 710 is used to obtain original multimedia data; image acquisition module 720 is used to obtain a reference image; image adjustment module 730 is used to adjust the image frames in the original multimedia data according to the display properties of the reference image to obtain target multimedia data and display the target multimedia data. The target multimedia data includes a sequence of image frames in the original multimedia data adjusted based on the display properties of the reference image.
可选地,还包括:特效元素生成模块,用于在获取参考图像之后,根据所述参考图像生成特效元素,在所述原始多媒体数据中叠加显示所述参考图像和特效元素。所述特效元素叠加在所述参考图像的设定位置显示并按照设定的方向移动;在所述特效元素在所述原始多媒体数据中的显示状态满足设定条件时,执行显示所述目标多媒体数据的步骤。Optionally, the method further includes: a special effect element generation module, which is used to generate a special effect element according to the reference image after acquiring the reference image, and to superimpose and display the reference image and the special effect element in the original multimedia data. The special effect element is superimposed on the reference image and displayed at a set position and moves in a set direction; when the display state of the special effect element in the original multimedia data meets the set condition, the step of displaying the target multimedia data is executed.
进一步地,所述根据所述参考图像生成特效元素,包括:根据所述参考图像中像素的色彩属性确定所述参考图像包含的色彩类别;生成与所述色彩类别表征的色彩对应的特效元素。Furthermore, generating a special effect element according to the reference image includes: determining a color category included in the reference image according to color attributes of pixels in the reference image; and generating a special effect element corresponding to a color represented by the color category.
进一步地,所述在所述原始多媒体数据中叠加显示所述参考图像和特效元素,包括:在所述原始多媒体数据的图像帧中叠加显示所述参考图像;在所述参考图像中叠加显示线条素材,所述线条素材在所述参考图像中沿第一方向移动;按照设定叠加形式将所述特效元素叠加到所述参考图像中的第一区域;将所述特效元素从所述参考图像的第一区域沿第二方向移动至所述参考图像对应的第二区域,调整所述特效元素在所述第二区域的显示状态。Furthermore, the superimposing and displaying the reference image and special effect element in the original multimedia data includes: superimposing and displaying the reference image in the image frame of the original multimedia data; superimposing and displaying line material in the reference image, wherein the line material moves along a first direction in the reference image; superimposing the special effect element to a first area in the reference image according to a set superimposition form; moving the special effect element from the first area of the reference image along a second direction to a second area corresponding to the reference image, and adjusting the display status of the special effect element in the second area.
进一步地,所述按照设定叠加形式将所述特效元素叠加到所述参考图像中的第一区域,包括:根据所述设定叠加形式确定所述特效元素对应的显示位置和亮度排列顺序;根据所述亮度排列顺序确定所述特效元素的显示顺序;按照所述显示位置和显示顺序,在所述原始多媒体数据的图像帧中的参考图像的第一区域中叠加显示所述特效元素。Furthermore, superimposing the special effect element to the first area in the reference image according to the set superimposition form includes: determining a display position and a brightness arrangement order corresponding to the special effect element according to the set superimposition form; determining a display order of the special effect elements according to the brightness arrangement order; and superimposing and displaying the special effect element in the first area of the reference image in the image frame of the original multimedia data according to the display position and display order.
可选地,图像调整模块720具体用于:对于所述参考图像和原始多媒体数据中的图像帧,分别根据所述显示属性确定图像对应的目标像素区间;对于每个所述目标像素区间,基于所述参考图像的像素的显示属性对所述原始多媒体数据中的图像帧进行属性匹配,融合每个所述目标像素区间对应的属性匹配结果,得到所述图像帧对应的多媒体数据处理结果;对于所述原始多媒体数据的图像帧中肤色区域中的像素,根据所述参考图像的肤色区域的显示属性调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果;对于所述原始多媒体数据中的图像帧,根据肤色权重融合所述肤色调整结果和多媒体数据处理结果,得到目标多媒体数据。所述肤色权重表征原始多媒体数据的图像帧中每个像素属于肤色区域的概率。Optionally, the image adjustment module 720 is specifically used to: for the reference image and the image frames in the original multimedia data, determine the target pixel intervals corresponding to the images according to the display attributes respectively; for each of the target pixel intervals, perform attribute matching on the image frames in the original multimedia data based on the display attributes of the pixels of the reference image, fuse the attribute matching results corresponding to each of the target pixel intervals, and obtain the multimedia data processing results corresponding to the image frames; for the pixels in the skin color area of the image frames of the original multimedia data, adjust the display attributes of the pixels according to the display attributes of the skin color area of the reference image, and obtain the skin color adjustment results corresponding to the image frames; for the image frames in the original multimedia data, fuse the skin color adjustment results and the multimedia data processing results according to the skin color weight, and obtain the target multimedia data. The skin color weight represents the probability that each pixel in the image frame of the original multimedia data belongs to the skin color area.
进一步地,所述对于所述原始多媒体数据的图像帧中肤色区域中的像素,根据所述参考图像的肤色区域的显示属性调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果,包括:确定所述原始多媒体数据的图像帧中每个像素属于肤色区域的概率,作为肤 色权重,根据所述肤色权重确定所述原始多媒体数据的图像帧中肤色区域包含的像素;根据所述原始多媒体数据的图像帧中肤色区域包含的像素的显示属性,确定所述原始多媒体数据的图像帧中肤色区域的显示属性均值;对于所述原始多媒体数据的图像帧中肤色区域中的像素,根据所述图像帧和参考图像分别在肤色区域的显示属性均值调整所述像素的显示属性,得到所述图像帧对应的肤色调整结果。Further, for a pixel in a skin color area in an image frame of the original multimedia data, adjusting the display attribute of the pixel according to the display attribute of the skin color area of the reference image to obtain a skin color adjustment result corresponding to the image frame includes: determining a probability that each pixel in the image frame of the original multimedia data belongs to the skin color area, as the skin color The method comprises the steps of: determining pixels included in a skin color area in an image frame of the original multimedia data according to the skin color weight; determining a mean value of display attributes of the skin color area in the image frame of the original multimedia data according to display attributes of the pixels included in the skin color area in the image frame of the original multimedia data; and adjusting the display attributes of the pixels in the skin color area in the image frame of the original multimedia data according to the mean values of display attributes of the image frame and a reference image in the skin color area, so as to obtain a skin color adjustment result corresponding to the image frame.
本公开实施例所提供的多媒体数据的处理装置可执行本公开任意实施例所提供的多媒体数据的处理方法,具备执行方法相应的功能模块和有益效果。The multimedia data processing device provided in the embodiments of the present disclosure can execute the multimedia data processing method provided in any embodiment of the present disclosure, and has the corresponding functional modules and beneficial effects of the execution method.
值得注意的是,上述装置所包括的各个单元和模块只是按照功能逻辑进行划分的,但并不局限于上述的划分,只要能够实现相应的功能即可;另外,各功能单元的具体名称也只是为了便于相互区分,并不用于限制本公开实施例的保护范围。It is worth noting that the various units and modules included in the above-mentioned device are only divided according to functional logic, but are not limited to the above-mentioned division, as long as the corresponding functions can be achieved; in addition, the specific names of the functional units are only for the convenience of distinguishing each other, and are not used to limit the protection scope of the embodiments of the present disclosure.
图8为本公开实施例所提供的一种电子设备的结构示意图。下面参考图8,其示出了适于用来实现本公开实施例的电子设备(例如图8中的终端设备或服务器)800的结构示意图。本公开实施例中的终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图8示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。FIG8 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present disclosure. Referring to FIG8 below, it shows a schematic diagram of the structure of an electronic device (e.g., a terminal device or server in FIG8 ) 800 suitable for implementing an embodiment of the present disclosure. The terminal device in the embodiment of the present disclosure may include, but is not limited to, mobile terminals such as mobile phones, laptop computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), vehicle-mounted terminals (e.g., vehicle-mounted navigation terminals), etc., and fixed terminals such as digital TVs, desktop computers, etc. The electronic device shown in FIG8 is merely an example and should not impose any limitations on the functions and scope of use of the embodiments of the present disclosure.
如图8所示,电子设备800可以包括处理装置(例如中央处理器、图形处理器等)801,其可以根据存储在只读存储器(ROM)802中的程序或者从存储装置808加载到随机访问存储器(RAM)803中的程序而执行各种适当的动作和处理。在RAM 803中,还存储有电子设备800操作所需的各种程序和数据。处理装置801、ROM 802以及RAM 803通过总线804彼此相连。编辑/输出(I/O)接口805也连接至总线804。As shown in FIG8 , the electronic device 800 may include a processing device (e.g., a central processing unit, a graphics processing unit, etc.) 801, which can perform various appropriate actions and processes according to a program stored in a read-only memory (ROM) 802 or a program loaded from a storage device 808 to a random access memory (RAM) 803. In the RAM 803, various programs and data required for the operation of the electronic device 800 are also stored. The processing device 801, the ROM 802, and the RAM 803 are connected to each other via a bus 804. An edit/output (I/O) interface 805 is also connected to the bus 804.
通常,以下装置可以连接至I/O接口805:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置806;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置807;包括例如磁带、硬盘等的存储装置808;以及通信装置809。通信装置809可以允许电子设备800与其他设备进行无线或有线通信以交换数据。虽然图8示出了具有各种装置的电子设备800,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。Typically, the following devices may be connected to the I/O interface 805: input devices 806 including, for example, a touch screen, a touchpad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; output devices 807 including, for example, a liquid crystal display (LCD), a speaker, a vibrator, etc.; storage devices 808 including, for example, a magnetic tape, a hard disk, etc.; and communication devices 809. The communication device 809 may allow the electronic device 800 to communicate wirelessly or wired with other devices to exchange data. Although FIG. 8 shows an electronic device 800 with various devices, it should be understood that it is not required to implement or have all the devices shown. More or fewer devices may be implemented or have alternatively.
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在非暂态计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置809从网络上被下载和安装,或者从存储装置808被安装,或者从ROM 802被安装。在该计算机程序被处理装置801执行时,执行 本公开实施例的方法中限定的上述功能。In particular, according to an embodiment of the present disclosure, the process described above with reference to the flowchart can be implemented as a computer software program. For example, an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a non-transitory computer-readable medium, and the computer program contains program code for executing the method shown in the flowchart. In such an embodiment, the computer program can be downloaded and installed from a network through a communication device 809, or installed from a storage device 808, or installed from a ROM 802. When the computer program is executed by the processing device 801, the execution The above functions are defined in the method of the embodiment of the present disclosure.
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的,而并不是用于对这些消息或信息的范围进行限制。The names of the messages or information exchanged between multiple devices in the embodiments of the present disclosure are only used for illustrative purposes and are not used to limit the scope of these messages or information.
本公开实施例提供的电子设备与上述实施例提供的多媒体数据的处理方法属于同一发明构思,未在本实施例中详尽描述的技术细节可参见上述实施例,并且本实施例与上述实施例具有相同的有益效果。The electronic device provided by the embodiment of the present disclosure and the method for processing multimedia data provided by the above embodiment belong to the same inventive concept. For technical details not fully described in this embodiment, reference can be made to the above embodiment, and this embodiment has the same beneficial effects as the above embodiment.
本公开实施例提供了一种计算机存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述实施例所提供的多媒体数据的处理方法。The embodiments of the present disclosure provide a computer storage medium on which a computer program is stored. When the program is executed by a processor, the method for processing multimedia data provided by the above embodiments is implemented.
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。It should be noted that the computer-readable medium disclosed above may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two. The computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination of the above. More specific examples of computer-readable storage media may include, but are not limited to: an electrical connection with one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above. In the present disclosure, a computer-readable storage medium may be any tangible medium containing or storing a program that may be used by or in combination with an instruction execution system, device or device. In the present disclosure, a computer-readable signal medium may include a data signal propagated in a baseband or as part of a carrier wave, in which a computer-readable program code is carried. This propagated data signal may take a variety of forms, including but not limited to an electromagnetic signal, an optical signal, or any suitable combination of the above. The computer readable signal medium may also be any computer readable medium other than a computer readable storage medium, which may send, propagate or transmit a program for use by or in conjunction with an instruction execution system, apparatus or device. The program code contained on the computer readable medium may be transmitted using any suitable medium, including but not limited to: wires, optical cables, RF (radio frequency), etc., or any suitable combination of the above.
在一些实施方式中,客户端、服务器可以利用诸如HTTP(HyperText Transfer Protocol,超文本传输协议)之类的任何当前已知或未来研发的网络协议进行通信,并且可以与任意形式或介质的数字数据通信(例如,通信网络)互连。通信网络的示例包括局域网(“LAN”),广域网(“WAN”),网际网(例如,互联网)以及端对端网络(例如,ad hoc端对端网络),以及任何当前已知或未来研发的网络。In some embodiments, the client and server may communicate using any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), an internet (e.g., the Internet), and a peer-to-peer network (e.g., an ad hoc peer-to-peer network), as well as any currently known or future developed network.
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。The computer-readable medium may be included in the electronic device, or may exist independently without being incorporated into the electronic device.
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:获取原始多媒体数据;获取参考图像;根据所述参考图像的显示属性调整所述原始多媒体数据中的图像帧,得到目标多媒体数据,显示所述目标多媒 体数据。所述目标多媒体数据包括所述原始多媒体数据中基于所述参考图像的显示属性调整后的图像帧序列。The computer-readable medium carries one or more programs. When the one or more programs are executed by the electronic device, the electronic device: obtains original multimedia data; obtains a reference image; adjusts the image frame in the original multimedia data according to the display attribute of the reference image to obtain target multimedia data, and displays the target multimedia data. The target multimedia data includes an image frame sequence in the original multimedia data that is adjusted based on the display attributes of the reference image.
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括但不限于面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for performing the operations of the present disclosure may be written in one or more programming languages or a combination thereof, including, but not limited to, object-oriented programming languages, such as Java, Smalltalk, C++, and conventional procedural programming languages, such as "C" or similar programming languages. The program code may be executed entirely on the user's computer, partially on the user's computer, as a separate software package, partially on the user's computer and partially on a remote computer, or entirely on a remote computer or server. In cases involving a remote computer, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (e.g., through the Internet using an Internet service provider).
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flow chart and block diagram in the accompanying drawings illustrate the possible architecture, function and operation of the system, method and computer program product according to various embodiments of the present disclosure. In this regard, each square box in the flow chart or block diagram can represent a module, a program segment or a part of a code, and the module, the program segment or a part of the code contains one or more executable instructions for realizing the specified logical function. It should also be noted that in some implementations as replacements, the functions marked in the square box can also occur in a sequence different from that marked in the accompanying drawings. For example, two square boxes represented in succession can actually be executed substantially in parallel, and they can sometimes be executed in the opposite order, depending on the functions involved. It should also be noted that each square box in the block diagram and/or flow chart, and the combination of the square boxes in the block diagram and/or flow chart can be implemented with a dedicated hardware-based system that performs a specified function or operation, or can be implemented with a combination of dedicated hardware and computer instructions.
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。单元的名称在某种情况下并不构成对该单元本身的限定。The units involved in the embodiments described in the present disclosure may be implemented by software or hardware. The name of a unit does not limit the unit itself in some cases.
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、片上系统(SOC)、复杂可编程逻辑设备(CPLD)等等。The functions described above herein may be performed at least in part by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on chips (SOCs), complex programmable logic devices (CPLDs), and the like.
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。 In the context of the present disclosure, a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, device, or equipment. A machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or equipment, or any suitable combination of the foregoing. A more specific example of a machine-readable storage medium may include an electrical connection based on one or more lines, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。The above description is only a preferred embodiment of the present disclosure and an explanation of the technical principles used. Those skilled in the art should understand that the scope of disclosure involved in the present disclosure is not limited to the technical solutions formed by a specific combination of the above technical features, but should also cover other technical solutions formed by any combination of the above technical features or their equivalent features without departing from the above disclosed concept. For example, the above features are replaced with the technical features with similar functions disclosed in the present disclosure (but not limited to) by each other to form a technical solution.
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。In addition, although each operation is described in a specific order, this should not be understood as requiring these operations to be performed in the specific order shown or in a sequential order. Under certain circumstances, multitasking and parallel processing may be advantageous. Similarly, although some specific implementation details are included in the above discussion, these should not be interpreted as limiting the scope of the present disclosure. Some features described in the context of a separate embodiment can also be implemented in a single embodiment in combination. On the contrary, the various features described in the context of a single embodiment can also be implemented in multiple embodiments individually or in any suitable sub-combination mode.
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。 Although the subject matter has been described in language specific to structural features and/or methodological logical actions, it should be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or actions described above. On the contrary, the specific features and actions described above are merely example forms of implementing the claims.
Claims (11)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202311415449.XA CN117425086A (en) | 2023-10-27 | 2023-10-27 | Multimedia data processing method and device, electronic equipment and storage medium |
| CN202311415449.X | 2023-10-27 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025087392A1 true WO2025087392A1 (en) | 2025-05-01 |
Family
ID=89529908
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2024/127406 Pending WO2025087392A1 (en) | 2023-10-27 | 2024-10-25 | Multimedia data processing method and apparatus, and electronic device and storage medium |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN117425086A (en) |
| WO (1) | WO2025087392A1 (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117425086A (en) * | 2023-10-27 | 2024-01-19 | 北京字跳网络技术有限公司 | Multimedia data processing method and device, electronic equipment and storage medium |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116703707A (en) * | 2023-05-29 | 2023-09-05 | 深圳数联天下智能科技有限公司 | Method for training skin color migration model, method for generating skin care image and related device |
| CN116723357A (en) * | 2023-07-07 | 2023-09-08 | 北京字跳网络技术有限公司 | Video special effect adding method, device, equipment, storage medium and program product |
| CN116758305A (en) * | 2023-05-23 | 2023-09-15 | 北京字跳网络技术有限公司 | Image color adjustment methods, devices, equipment, readable storage media and products |
| CN117425086A (en) * | 2023-10-27 | 2024-01-19 | 北京字跳网络技术有限公司 | Multimedia data processing method and device, electronic equipment and storage medium |
-
2023
- 2023-10-27 CN CN202311415449.XA patent/CN117425086A/en active Pending
-
2024
- 2024-10-25 WO PCT/CN2024/127406 patent/WO2025087392A1/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116758305A (en) * | 2023-05-23 | 2023-09-15 | 北京字跳网络技术有限公司 | Image color adjustment methods, devices, equipment, readable storage media and products |
| CN116703707A (en) * | 2023-05-29 | 2023-09-05 | 深圳数联天下智能科技有限公司 | Method for training skin color migration model, method for generating skin care image and related device |
| CN116723357A (en) * | 2023-07-07 | 2023-09-08 | 北京字跳网络技术有限公司 | Video special effect adding method, device, equipment, storage medium and program product |
| CN117425086A (en) * | 2023-10-27 | 2024-01-19 | 北京字跳网络技术有限公司 | Multimedia data processing method and device, electronic equipment and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| CN117425086A (en) | 2024-01-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN112561840B (en) | Video clipping method and device, storage medium and electronic equipment | |
| WO2022068451A1 (en) | Style image generation method and apparatus, model training method and apparatus, device, and medium | |
| US20240112299A1 (en) | Video cropping method and apparatus, storage medium and electronic device | |
| WO2022142875A1 (en) | Image processing method and apparatus, electronic device, and storage medium | |
| WO2022171036A1 (en) | Video target tracking method, video target tracking apparatus, storage medium, and electronic device | |
| CN112348910B (en) | Method, device, apparatus and computer readable medium for acquiring images | |
| CN110689478A (en) | Image stylization processing method, device, electronic device and readable medium | |
| WO2024037556A1 (en) | Image processing method and apparatus, and device and storage medium | |
| CN112034984A (en) | A virtual model processing method, device, electronic device and storage medium | |
| WO2024240222A1 (en) | Image stylization processing method and apparatus, device, storage medium and program product | |
| WO2023078284A1 (en) | Image rendering method and apparatus, device, storage medium, and program product | |
| CN111754613A (en) | Image decoration method, apparatus, computer readable medium and electronic device | |
| WO2025087392A1 (en) | Multimedia data processing method and apparatus, and electronic device and storage medium | |
| WO2022233223A1 (en) | Image splicing method and apparatus, and device and medium | |
| CN112488947B (en) | Model training and image processing method, device, equipment and computer-readable medium | |
| CN110399802B (en) | Method, apparatus, medium, and electronic device for processing eye brightness of face image | |
| CN114372974B (en) | Image detection method, device, equipment and storage medium | |
| WO2025087280A1 (en) | Image processing method and apparatus, electronic device, and storage medium | |
| WO2025061019A1 (en) | Image presentation method and apparatus, and storage medium and electronic device | |
| WO2024131503A1 (en) | Special-effect image generation method and apparatus, and device and storage medium | |
| CN118840453A (en) | Text color determining method, text color determining device, text color determining medium, electronic equipment and program product | |
| US20230360286A1 (en) | Image processing method and apparatus, electronic device and storage medium | |
| JP7752184B2 (en) | Image brightness adjustment method, device, electronic device, and medium | |
| CN114170341B (en) | Image processing method, device, equipment and medium | |
| WO2024131652A1 (en) | Special effect processing method and apparatus, and electronic device and storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24881752 Country of ref document: EP Kind code of ref document: A1 |