WO2019196941A1 - Configuration de transformation implicite adaptative - Google Patents
Configuration de transformation implicite adaptative Download PDFInfo
- Publication number
- WO2019196941A1 WO2019196941A1 PCT/CN2019/082537 CN2019082537W WO2019196941A1 WO 2019196941 A1 WO2019196941 A1 WO 2019196941A1 CN 2019082537 W CN2019082537 W CN 2019082537W WO 2019196941 A1 WO2019196941 A1 WO 2019196941A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- intra prediction
- transform
- mode
- current block
- modes
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/12—Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
Definitions
- the present disclosure relates generally tovideo processing.
- the present disclosure relates to implicit signaling of transform settings based on intra prediction settings.
- High-Efficiency Video Coding is a new international video coding standard developed by the Joint Collaborative Team on Video Coding (JCT-VC) .
- JCT-VC Joint Collaborative Team on Video Coding
- HEVC is based onthe hybrid block-based motion-compensated Discrete Cosine Transform (DCT) coding architecture.
- the basic unit for compression termed coding unit (CU) , is a 2Nx2N square block, and each CU can be recursively split into four smaller CUs until the predefined minimum size is reached.
- Each CU contains one or multiple prediction units (PUs) . For each PU, either Intra Prediction or Inter Prediction can be used.
- PUs prediction units
- HEVC Discrete Cosine Transform type II (DCT-II) as its core transform because it has a strong "energy compaction" property.
- DCT-II Discrete Cosine Transform type II
- a video encoder receives a block of pixels of a video picture to be coded as a current block.
- the video encoder performs an intra prediction operation for the current block according to an intra prediction settingand generating prediction residuals of the current block.
- the intra prediction setting selects a current intra prediction mode from a plurality of intra prediction modes for the current block.
- the video encoder maps the current intra prediction mode to a target transform mode.
- the video encoder encodes the current block in a bitstream by performing a transform operation on the prediction residuals of the current block according to the target transform mode.
- a video decoder that implicitly decides transform settings based on intra prediction settings.
- a video decoder receives a bitstream comprising coded prediction residuals of a current block of a video picture.
- the video decoder determines a current intra prediction mode from a plurality of intra prediction modes for the current block according to an intra prediction setting and maps the current intra prediction mode to a target transform mode.
- the video decoder decodes the current block by performing an inverse transform operation on the coded prediction residuals of the current blockaccording to the target transform mode to generate prediction residual of the current block and by reconstructing the current block based on the prediction residual and a set of prediction pixels of the current block, in which set of prediction pixels of the current block is generated by performing an intra prediction operation for the current block according to thecurrent intra prediction mode.
- a collection of intra prediction modes are assigned with a particular transform mode, such that the target transform mode is the particular transform mode when the current intra prediction mode is in the collection of intra prediction modes.
- the collection of intra prediction modes may be defined by a fixed expression, such that different blocks with different block sizes have different fixed expression for defining the collection of intra prediction modes, or such that different fixed expressions are used to assign intra prediction modes in different numerical ranges with transform modes.
- intra prediction modes in a most probable mode (MPM) list of the current block are assigned with a particular transform mode.
- an intra prediction mode indicative of uniform block texture, such as DC mode is mapped to a particular transform mode.
- FIG. 1 conceptually illustrates implicit determination of transform settings based on prediction settings in a video encoder.
- FIG. 2 conceptually illustrates implicit determination of transform settings based on intra prediction settings in a video decoder.
- FIG. 3A illustrates an example Adaptive Multiple Transform (AMT) scheme.
- AMT Adaptive Multiple Transform
- FIG. 3B illustrates an example Multiple Transform Selection (MTS) scheme.
- MTS Multiple Transform Selection
- FIG. 4 illustrates the intra prediction modes for coding a block of pixels.
- FIG. 5 illustrates an example video encoder that may implement implicit signaling of transform settings based on intra prediction settings.
- FIG. 6 illustrates portions of the video encoder that implement implicit derivation of transform settings based on intra prediction settings.
- FIG. 7 conceptually illustrates a video encoding process that implicitly signals transform settings based on intra prediction settings.
- FIG. 8 illustrates an example video decoder that may implicitly signal transform settings based on intra prediction settings.
- FIG. 9 illustrates portions of the video decoder that implement implicit derivation of transform settings based on intra prediction settings.
- FIG. 10 conceptually illustrates a video decoding process that implicitly decidestransform settings based on intra prediction settings.
- FIG. 11 conceptually illustrates an electronic system with which some embodiments of the present disclosure are implemented.
- Some embodiments of the disclosure provide a video coder that implicitly signal transform setting (s) according to a set of predefined rules.
- the video coder may derive a transform mode and/or other transform settings based on intra prediction settings that are explicitly signaled in a bitstream.
- the transform mode may specify a vertical transform type and horizontal transform type that are selected multiple different transform types.
- FIG. 1 conceptually illustrates implicit determination of transform settings based on prediction settings in a video encoder 100.
- the video encoder 100 receives raw pixels from a video source 110. The raw pixels belong to video pictures of a video sequence.
- the video encoder 100 divides the raw pixels belonging to a video picture into pixel blocks (e.g., CUs) to be encoded into a bitstream 190 as syntax elements.
- pixel blocks e.g., CUs
- the video encoder 100 includes a transform module 120, a set of transform settings 125, an intra prediction module 130, a collection of intra prediction settings 135, and an implicitmapping module 140.
- the modules 120-140 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device or electronic apparatus.
- the modules 120-140 are modules of hardware circuits implemented by one or more integrated circuits (ICs) of an electronic apparatus. Though the modules 120-140 are illustrated as being separate modules, some of the modules can be combined into a single module.
- theintra prediction module 130 When coding a block of pixels (the current block) , theintra prediction module 130 generates a set of prediction pixels by referencing pixels neighboring the current block in different directions. The generated prediction pixel is subtracted from raw pixel values of the current block to produce a set of residuals, and the residuals are transformed into transform coefficients by the transform module 120. The transform coefficients are then entropy coded as syntax element of the bitstream 190.
- the transform module 120 may include multiple different transform engines, such as transform engines for DCT-II, DST-VII, DCT-VIII, DST-I, and DCT-V transform types or any subset of the above types.
- the transform settings 125 select which transform engine is used to transform the residual values into transform coefficients for the current block. In some embodiments, the transform settings 125 specify a horizontal transform type and a vertical transform type.
- the intra prediction module 130 performs intra prediction by referencing pixels within the current video picture.
- the operations of the intra prediction module 130 are controlled by the intra prediction settings 135, which specifies an intra prediction mode that is selected from DC mode, planar mode, and multiple different directional/angular modes. Intra prediction modes are described in greater detail in Section II below.
- the intra prediction settings 135 may be determined by a rate distortion control engine (not illustrated) of the video encoder 100.
- the intra prediction settings 135 may be explicitly coded as syntax elements in the bitstream 190.
- the transform settings 125 are derived or mapped from the intra prediction settings 135 and not coded in the bitstream as syntax elements. In other words, the intra prediction settings 135 are explicitly signaled while the transform settings 125 are implicitly signaled.
- the implicit mapping module 140 maps the intra prediction settings 135 into transform settings 125.
- the implicit mapping module 140 defines a collection of intra prediction modesaccording to a predetermine criteria.
- the defined collection of intra prediction modes are assigned with or mapped to a transform group or a transform mode according to an Adaptive Multiple Transform (AMT) Scheme or a multiple transform selection (MTS) scheme.
- AMT specifies the horizontal and vertical transform types for different transform groups and different transform modes.
- MTS specifies the horizontal and vertical transform types for different transform modes.
- AMT and MTS schemes are described in greater detail in Section I below.
- the implicit mapping module 140 determines whether the particular intra prediction mode falls within the defined collection of intra prediction modes. If so, the implicit mapping module 140 provides a horizontal type and a vertical transform type based on the transform group and/or transform mode that the defined collection of intra prediction modes is assigned with or mapped to. In some embodiments, the implicit mapping module 140 defines multiple collections of intra prediction modes and maps each defined collection of intra prediction modes to a different transform group or transform mode. In some embodiments, the mapping of intra prediction settings to transform settings is based on predefined functions. In some embodiments, the mapping of intra prediction settings to transform settings is further based on other properties of the current block, such as the size/width/height of the current block.
- the implicit mapping module 140 defines the collection of intra prediction modes based on a list of most probable modes (MPM) that is identified based on information (e.g., prediction modes) of the blocks neighboring the current block.
- MPM most probable modes
- the MPM are described in Section III below.
- the mapping of intra prediction settings to transform settings is described in greater detail in Section IV below.
- FIG. 2 conceptually illustrates implicit determination of transform settings based on intra prediction settings in a video decoder 200.
- the figure illustrates the video decoder 200 receiving syntax elements of a bitstream 210.
- the video decoder 200 decodes the syntax elements to reconstruct video pictures of a video sequence for transmission or display.
- the video decoder 200 includes an inverse transform module 220, a set of transform settings 225, an intra prediction module 230, a collection of intra prediction settings 235, and an implicit mapping module 240.
- the modules 220-240 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device or electronic apparatus.
- the modules 220-240 are modules of hardware circuits implemented by one or more integrated circuits (ICs) of an electronic apparatus. Though the modules 220-240 are illustrated as being separate modules, some of the modules can be combined into a single module.
- the intra prediction module 230 When coding a block of pixels (the current block) , the intra prediction module 230 generates a set of prediction pixels by referencing pixels neighboring the current block in different directions. The generated prediction pixel is added with a set of residual values that are produced by the inverse transform module 220 based on transform coefficients. The transform coefficients are entropy decoded from syntax elements of the bitstream 210.
- the inverse transform module 220 may include multiple different transform engines, such as transform engines for DCT-II, DST-VII, DCT-VIII, DST-I, and DCT-V transform types, or any subset of the above types.
- the transform settings 125 select which transform engine is used to inverse transform the transform coefficients into residual values for the current block.
- the transform settings 225 specifies a horizontal transform type and a vertical transform type.
- the intra prediction module 230 performs intra prediction by referencing pixels within the current video picture.
- the operations of the intra prediction module 230 are controlled by the intra prediction settings 235, which specifies an intra prediction mode that is selected from DC mode, planar mode, and multiple different directional/angular modes. Intra prediction modes are described in greater detail in Section II below.
- the intra prediction settings 235 are based on syntax elements parsed from the bitstream.
- the transform settings 225 are derived or mapped from the intra prediction settings 235 and not coded in the bitstream as syntax elements. In other words, the intra prediction settings 235 are explicitly signaled while the transform settings 225 are implicitly signaled.
- the implicit mapping module 240 maps the intra prediction settings 235 into transform settings 225.
- the implicit mapping module 240 defines a collection of intra prediction modes according to a predetermine criteria.
- AMT Adaptive Multiple Transform
- MTS multiple transform selection
- the implicit mapping module 240 determines whether the particular intra prediction mode falls within the defined collection of intra prediction modes. If so, the implicit mapping module 240 provides a horizontal type and a vertical transform type based on the transform group and/or transform mode that the defined collection of intra prediction mode is assigned with or mapped to. In some embodiments, the implicit mapping module 240 defines multiple collectionsof intra prediction modes and maps each defined collection of intra prediction modes to a different transform group or transform mode. In some embodiments, the mapping of intra prediction settings to transform settings is based on predefined functions. In some embodiments, the mapping of intra prediction settings to transform settings is further based on other properties of the current block, such as the size/width/height of the current block.
- the implicit mapping module 240 defines the collection of intra prediction modes based on a list of most probable modes (MPM) that is identified based on information (e.g., prediction modes) of the blocks neighboring the current block.
- MPM most probable modes
- the MPM are described in Section III below.
- the mapping of intra prediction settings to transform settings is described in greater detail in Section IV below.
- an Adaptive Multiple Transform (AMT) scheme is used to define the transform settings for residual coding for both intra and inter coded blocks.
- AMT scheme uses transform modes selected from the DCT/DST families or transform types, including Discrete Sine Transform type VII (DST-VII) , Discrete Cosine Transform type VIII (DCT-VIII) , Discrete Sine Transform type I (DST-I) , and Discrete Cosine Transform Type V (DCT-V) .
- multiple transform modes used in AMT scheme are divided into Transform Group 1 and Transform Group 2.
- the number of transform modes in Transform Group 1 is denoted as A, where A is a positive integer.
- the number of transform modes in Transform Group 2 is denoted as B, where B is a positive integer.
- a plus B is equal to the total number of transform modes.
- a transform mode may assign one transform type for horizontal transform and one transform type for vertical transform.
- FIG. 3A illustrates an example AMT scheme 300A.
- the example AMT scheme 300 Ause onlytransform types DCT-II and DST-VII, and there is one transform mode in Transform Group 1 and there are three transform modes in Transform Group 2.
- the one transform mode (mode 1-1) defines DCT-II for both horizontal transform type and vertical transform type.
- the first transform mode (mode 2-1) defines DST-VII for both horizontal transform and vertical transform
- the second transform mode (mode 2-2) defines DCT-II for horizontal transform and DST-VII for vertical transform
- the third transform mode (mode 2-3) defines DST-VII for horizontal transform and DCT-II for vertical transform.
- the transform modes in the Group 1 and Group 2 are different from each other.
- An AMT flag (also referred to as transformflag) issignaled to indicate a particular transform group. If the number of transform modes in the particular transform group is larger than 1, an additional transform index may besignaled to indicate one transform mode.
- a transform group is implicitly assigned or selected according to predetermined criteria such that the AMT flag is not signaled.
- the transform mode is implicitly assigned or selected according to the predetermined criteria such that neither the AMT flag nor the AMT index is signaled.
- the predetermined criteria is based on the applied Intra prediction mode of the current block.
- MTS Multiple Transform Selection
- DCT/DST families or transform types including Discrete Sine Transform type VII (DST-VII) and/or Discrete Cosine Transform type VIII (DCT-VIII) .
- multiple transform modes used in MTS are indicated with an index.
- a transform mode may assign one transform type for horizontal transform and one transform type for vertical transform.
- FIG. 3B illustrates an example MTSscheme 300B.
- the example MTS300B uses only transform types DST-VII and DCT-VIII.
- Adefault mode (not illustrated) defines DCT-II for both horizontal and vertical transform.
- the MTS scheme 300B defines four additional transform modes 1 through 4.
- Transform mode 1 defines DST-VII for both horizontal transform type and vertical transform type.
- Transform mode 2 defines DCT-VIII for horizontal transform type and DST-VII for vertical transform type.
- Transform mode 3 defines DST-VII for horizontal transform type and DCT-VIIIfor vertical transform type.
- Transform mode 4 defines DCT-VIII for both horizontal transform and vertical transform.
- the transform modes are different from each other.
- An MTS index is signaled to indicate a particular transform mode.
- a transform mode is implicitly assigned or selected according to predetermined criteria such that the MTS index is not signaled.
- the predetermined criteria is based on the applied Intra prediction mode of the current block.
- intra prediction in addition to DC and planar modes.
- intra prediction in order to improve intra prediction and to capture finer edge directions presented in natural videos, intra prediction is defined to have 65 directional modes, in addition to DC and planar modes.
- the denser directional intra prediction modes can be applied to all block sizes and both luma and chroma intra predictions.
- FIG. 4 illustrates the 67intra prediction modes for coding a block of pixels.
- intra prediction mode 0 corresponds to planar mode
- intra prediction mode 1 corresponds to DC mode
- intra prediction modes 2-66 correspond to angularintra prediction modes, or directional modes.
- mode 2 corresponds to bottom-left direction
- mode 18 corresponds to Horizontal or left direction
- mode 34 corresponds to Diagonal or top-left direction
- mode 50 corresponds to Vertical or top direction
- mode 66 corresponds to Vertical Diagonal or top-right direction.
- a mode-dependent transform candidate selection process is used to account for different residual statistics of different intra prediction modes.
- each of the intra prediction modes are assigned a set of candidate transforms, i.e., the set of candidate transforms are selected based on the intra prediction mode.
- a horizontal transform and a vertical transform are selected from the set of candidate transforms assigned to anintra prediction mode.
- Table 1 lists three pre-defined sets of candidate transforms, each set of candidate transforms may serve as a vertical transform set or a horizontal transform set.
- Table 2 lists 67 intra prediction modes and their assigned sets of candidate transforms.
- the video coder uses table 2 to identify the transform setbased on the particular intra prediction mode.
- the video coder uses a transform index (or AMT index) to select one transform mode from the vertical transform set and one transform mode from the horizontal transform set.
- the transform index may be explicitly coded in a bitstream as a syntax element.
- Table 1 Three pre-defined transform candidate sets
- Table 2 Selected (H) foral and (V) ertical transform sets for each intra prediction mode
- a list of Most Probable Modes is used to encode intra prediction mode for Luma components of a block.
- the list of MPMs is a list of candidate modes that is generated based on the intra predictionmodes used by neighboring blocks (or neighboring modes) of the current block.
- truncated unary coding may be used to indicate the current intra prediction mode if the current intra prediction mode is in the MPM list. If the current intra prediction modeis not in the MPM list, fixed length code or truncated binarization may be used to indicate the current intra prediction mode.
- intra prediction modes included in an MPM list may be classified into three groups: neighboring intra modes, derived intra modes, and default intra modes.
- intra prediction modesoffive neighboring blocks are used to form the MPM list.
- the locations of the 5 neighboring modes includes left (L) , above (A) , below-left (BL) , above-right (AR) , and above-left (AL) .
- An initial MPM list is formed by inserting intra prediction modes from the five neighboring modes, the planar mode, and the DC modes into the MPM list.
- a pruning process is used to remove duplicated intra predictionmodes so that only unique modes are included in the MPM list.
- the order of the intra prediction modes in the initial MPM list may be left, above, planar, DC, below-left, above-right, and then above-left.
- one or more derived intra modes may be added.
- a derived intra mode may be obtained by adding -1 or +1 to an angular directional mode (an intra prediction mode that is not DC or planar) that is already included in the MPM list. Such an additional derived mode may not be generated from a non-angular intra predictionmodes (DC or planar) .
- the MPM list is still not complete, one or more default modes are added to the MPM list according to the following order: vertical, horizontal, mode 2, and diagonal mode. As a result of this process, a unique list of 6 MPM candidate modes is generated.
- the coding for selection of the remaining 61 intra prediction modesthat are not in the MPM list (non-MPMs) is as follows.
- the 61 non-MPMs are divided into two mode sets: a selected mode set and a non-selected mode set.
- the selected mode set includes 16 modes; the remaining 45 modes are assigned to the non-selected modeset.
- the mode set (set of selectedmodes or set of non-selected modes) that includes the current mode (i.e., the intra prediction modeselected to code the current block) may be indicated in the bitstream with a flag. If the current mode is within the selected modeset, the current mode is signaled by a 4-bit fixed-length code. If the current mode is from the non-selected mode set, the current mode is signaled by a truncated binary code.
- the selected modeset is generated by sub-sampling the 61 non-MPM modes according to the following:
- Selected modeset ⁇ 0, 4, 8, 12, 16, 20 ... 60 ⁇
- Non-selected modeset ⁇ 1, 2, 3, 5, 6, 7, 9, 10 ... 59 ⁇
- the video coder forgoes explicit signaling of transform settings and instead uses implicit indications based on intra prediction settings, including planar mode, DC mode, and angular directional modes.
- the video coder may assign different transform settings to adjacent angular directional modes, e.g., based on similarity of predictors between adjacent directional modes.
- the video coder may also map planar mode and DC mode to transform settings depending on texture characteristic.
- DC mode is selected for intra prediction and may be implicitly assigned with or mapped to a default transform type such as DCT-II.
- DCT-II transform type
- both Intra prediction mode and transform mode are signaled at the same time using the Intra prediction mode coding.
- a video encoder testsdifferent Intra prediction modes to find the most suitable intra prediction mode, the corresponding choice for transform settings can be determined at the same time without extra coding pass and complexity.
- predetermined criteria is used to define a collection of Intra prediction modes, which areimplicitly assigned withor mapped to one particular transform group or one particular transform mode.
- the collection of intra prediction modes specifies the indices of the intra prediction modes that satisfy the predetermined criteria.
- the collection of intrapredictionmodes may be specified by a table with fixed entries. The table may containentries of different intra prediction modes, e.g., ⁇ 2, Horizontal, Diagonal, Vertical, Vertical diagonal ⁇ , or any other subset of these five prediction modes.
- the defined collection ofintra prediction modes may be implicitly assigned with or mapped toone transform group, such as Transform Group 1.
- the predetermined collection of intrapredictionmodes may also be implicitly assigned with one transform mode, such as DCT-II for both horizontal transform and vertical transform.
- the Intra prediction modes that are not in the defined collection of intra prediction modes may be implicitly assigned with another transform group, such as Transform Group 2 or another transform mode, such as DST-VII for both horizontal transform and vertical transform.
- the defined collection of intrapredictionmodes may be specified by a set of fixed entries of intra prediction modes such as ⁇ 2, Horizontal, Diagonal, Vertical, Vertical Diagonal ⁇ or any other subset of these five directional modes ⁇ offset, where offset can be predetermined or adaptively decided.
- intra prediction modes33 and 35 are implicitly assigned with one transform group (such as Transform Group 1) or one transform mode (such as DCT-II for both horizontal transform and vertical transform) while other Intra prediction modes are implicitly assigned with another transform group (such as Transform Group 2) or another transform mode (such as DST-VII for both horizontal transform and vertical transform) .
- one transform group such as Transform Group 1
- one transform mode such as DCT-II for both horizontal transform and vertical transform
- another transform group such as Transform Group 2
- another transform mode such as DST-VII for both horizontal transform and vertical transform
- the intra prediction modes 3, 5, 7, ..., 65 are then implicitly assigned with one transform groupor one transform mode while the other Intra prediction modes are implicitly assigned with another transform groupor another transform mode.
- Intra prediction modes 0, 2, 4, 6, ... 32, 35, 37, 39, ... 63, 65 are then implicitly assigned with one transform groupor one transform mode while the other intra prediction modes are implicitly assigned with another transform groupor another transform mode.
- the horizontal transform type for the intra prediction modes in the defined collection of intraprediction modes are the same as the vertical transform type for the intra prediction modes that are not in the defined collection of intraprediction modes, while the vertical transform type for the intra prediction modes in the define collection of intraprediction modes are the same as the horizontal transform type for the intra prediction modes that are not in the defined collection of intraprediction modes.
- the horizontal transform type may be DST-VII and the vertical transform type may be DCT-II
- the horizontal transform type may be DCT-II and the vertical transform type may be DST-VII.
- blocks with different block sizes may use different equations or expressions for defining the collection of intraprediction modes that are implicitly assigned with transform groups or transform modes.
- blocks applying intra prediction modes that are in different numerical ranges may use different equations or expressions for defining the collection of intraprediction modes that are implicitly assigned with transform groups or transform modes.
- the fixed equation may be ⁇ 2, Horizontal, Diagonal, Vertical, vertical diagonal ⁇ or any other subset of these five intra prediction modes ⁇ offset, where offset can be predetermined or adaptively decided.
- different fixed equations or expressions are used to assign intra prediction modes in different numerical ranges to transform modes or transform groups.
- the defined collection of intra prediction modesthat are implicitly assigned with a transform group or a transform mode may be specified by texture characteristic.
- a block having uniform texture may be implicitly assigned one transform group (e.g., Transform Group 1) or one transform mode (e.g., DCT-II)while a block not having uniform texture may be implicitly assigned another transform group (e.g., Transform Group 2) or another transform mode (e.g., DST-VII) .
- a video coder may determine texture uniformity by determining whether the applied Intra prediction mode of the block is DC mode.
- DC mode is the applied Intra prediction mode
- the block texture is assumed to be uniform, and the block is implicitly assigned Transform Group 1 or DCT-II, i.e., DC mode is implicitly assigned with Transform Group 1 or DCT-II.
- the applied intra prediction mode of the block is planar mode
- the block texture is assumed to be more complex, and the block is implicitly assigned Transform Group 2 or DST-VII, i.e., planar mode is implicitly assigned with Transform Group 2 or DST-VII.
- the defined collection of intra prediction modesthat are implicitly assigned with a transform group or a transform mode may be specified based on intra prediction modes of neighboring blocks.
- the intra prediction modes inthe MPM list of the current clock may be implicitly assigned with one transform group (e.g., Transform Group 1) or one transform mode (e.g., DCT-II for both horizontal transform and vertical transform) while the intra prediction modes that are not in the MPM list may be implicitly assigned with another transform group (e.g., Transform Group 2) or another transform mode (e.g., DST-VII for both horizontal transform and vertical transform) .
- Transform Group 1 e.g., Transform Group 1
- one transform mode e.g., DCT-II for both horizontal transform and vertical transform
- another transform group e.g., Transform Group 2
- another transform mode e.g., DST-VII for both horizontal transform and vertical transform
- theprediction modes ofthe left neighboring block or of the left-extended neighboring block may be implicitly assigned with one transform groupor one transform mode.
- the prediction modes from above neighboring block or above-extended neighboring block may be implicitly assigned with another transform groupor another transform mode.
- the prediction modes from above-left neighboring block may be implicitly assigned with another transform groupor another transform mode.
- the transform types in one transform group may be changed based on block size.
- a transform mode in Transform Group 1 and a transform mode in Transform Group 2 may be swapped based on certain criteria. For example, when the block size is smaller than a particular threshold, a transform mode in Transform Group 2 is exchanged with a transform mode in Transform Group 1.
- the only transform mode (mode 1-1) in Transform Group 1 is to be changed to DST-VII for both horizontal transform and vertical transform and the first transform mode (mode 2-1) in Transform Group 2 is to be changed to DCT-II for both horizontal transform and vertical transform.
- the combination of the transform type for horizontal transform and the transform type for vertical transform may be changed based on the applied intra prediction mode, such as whether the applied intra prediction mode is an odd (or even) indexed angular mode.
- the applied intra prediction mode is an odd number
- the second transform mode (mode 2-2) may be changed to DST-VII for horizontal transform type and DCT-II for vertical transform type
- the third transform mode (mode 2-3) may be changed to be DCT-II for horizontal transform type and DST-VII for vertical transform type.
- FIG. 5 illustrates an example video encoder 500 that may implement implicit signaling of transform settings based on intra prediction settings.
- the video encoder 500 may implement the video encoder 100 of FIG. 1.
- the video encoder 500 receives input video signal from a video source 505 and encodes the signal into bitstream 595.
- the video encoder 500 has several components or modules for encoding the signal from the video source 505, including a transform module 510, a quantization module 511, an inverse quantization module 514, an inverse transform module 515, an intra-picture estimation module 520, an intra prediction module 525, a motion compensation module 530, a motion estimation module 535, an in-loop filter 545, a reconstructed picture buffer 550, a MV buffer 565, and a MV prediction module 575, and an entropy encoder 590.
- the motion compensation module 530 and the motion estimation module 535 are part of an inter-prediction module 540.
- the modules 510 –590 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device or electronic apparatus. In some embodiments, the modules 510 –590 are modules of hardware circuits implemented by one or more integrated circuits (ICs) of an electronic apparatus. Though the modules 510 –590 are illustrated as being separate modules, some of the modules can be combined into a single module.
- the video source 505 provides a raw video signal that presents pixel data of each video frame without compression.
- a subtractor 508 computes the difference between the raw video pixel data of the video source 505 and the predicted pixel data 513 from the motion compensation module 530 or intra prediction module 525.
- the transform module 510 converts the difference (or the residual pixel data or residual signal 509) into transform coefficients (e.g., by performing Discrete Cosine Transform, or DCT) .
- the quantization module 511 quantizes the transform coefficients into quantized data (or quantized coefficients) 512, which is encoded into the bitstream 595 by the entropy encoder 590.
- the inverse quantization module 514 de-quantizes the quantized data (or quantized coefficients) 512 to obtain transform coefficients, and the inverse transform module 515 performs inverse transform on the transform coefficients to produce reconstructed residual 519.
- the reconstructed residual 519 is added with the predicted pixel data 513 to produce reconstructed pixel data 517.
- the reconstructed pixel data 517 is temporarily stored in a line buffer (not illustrated) for intra-picture prediction and spatial MV prediction.
- the reconstructed pixels are filtered by the in-loop filter 545 and stored in the reconstructed picture buffer 550.
- the reconstructed picture buffer 550 is a storage external to the video encoder 500.
- the reconstructed picture buffer 550 is a storage internal to the video encoder 500.
- the intra-picture estimation module 520 performs intra prediction based on the reconstructed pixel data 517 to produce intra prediction data.
- the intra prediction data is provided to the entropy encoder 590 to be encoded into bitstream 595.
- the intra prediction data is also used by the intra prediction module 525 to produce the predicted pixel data 513.
- the motion estimation module 535 performs inter-prediction by producing MVs to reference pixel data of previously decoded frames stored in the reconstructed picture buffer 550. These MVs are provided to the motion compensation module 530 to produce predicted pixel data.
- the video encoder 500 uses MV prediction to generate predicted MVs, and the difference between the MVs used for motion compensation and the predicted MVs is encoded as residual motion data and stored in the bitstream 595.
- the MV prediction module 575 generates the predicted MVs based on reference MVs that were generated for encoding previously video frames, i.e., the motion compensation MVs that were used to perform motion compensation.
- the MV prediction module 575 retrieves reference MVs from previous video frames from the MV buffer 565.
- the video encoder 500 stores the MVs generated for the current video frame in the MV buffer 565 as reference MVs for generating predicted MVs.
- the MV prediction module 575 uses the reference MVs to create the predicted MVs.
- the predicted MVs can be computed by spatial MV prediction or temporal MV prediction.
- the difference between the predicted MVs and the motion compensation MVs (MC MVs) of the current frame (residual motion data) are encoded into the bitstream 595 by the entropy encoder 590.
- the entropy encoder 590 encodes various parameters and data into the bitstream 595 by using entropy-coding techniques such as context-adaptive binary arithmetic coding (CABAC) or Huffman encoding.
- CABAC context-adaptive binary arithmetic coding
- the entropy encoder 590 encodes various header elements, flags, along with the quantized transform coefficients 512, and the residual motion data as syntax elements into the bitstream 595.
- the bitstream 595 is in turn stored in a storage device or transmitted to a decoder over a communications medium such as a network.
- the in-loop filter 545 performs filtering or smoothing operations on the reconstructed pixel data 517 to reduce the artifacts of coding, particularly at boundaries of pixel blocks.
- the filtering operation performed includes sample adaptive offset (SAO) .
- the filtering operations include adaptive loop filter (ALF) .
- FIG. 6 illustrates portions of the video encoder 500 that implement implicit derivation of transform settings based on intra prediction settings.
- intra prediction settings 600 control the operations of the intra prediction module 525.
- the intra prediction settings specify an intra prediction mode that is selected from DC mode, planar mode, and multiple different directional/angular modes. Intra prediction modes are described in Section II above.
- Animplicit mapping module 610 receives the intra prediction settings 600 and maps the received intra prediction settings to transform settings for the transform module 510.
- the transform settings may include a transform mode or transform types for a vertical transform and a horizontal transform.
- the transform types may be selected from various versions of DCT or DST such as DCT-II, DST-VII, DCT-VIII, DST-I, and DCT-V or any subset of the above types.
- the transform module 510 in turn performs transform on the prediction residuals of the block based on the derived transform settings.
- the entropy encoder 590 also receives the intra prediction settings 600 and encodes at least some of the prediction settings 600 as syntax element in the bitstream 595.
- the intra prediction settings 600 may be implicit signaled (i.e., not encoded in the bitstream) and can be implicitly derived by video decoders.
- the derived transform settings are not encoded in the bitstream.
- the mapping of the transform settings from the intra prediction settings is described in Section IV above.
- FIG. 7 conceptually illustrates a video encoding process 700 that implicitly signals transform settings based on intra prediction settings.
- one or more processing units e.g., a processor
- a computing device implementing the video encoder 500 performs the process 700 by executing instructions stored in a computer readable medium.
- an electronic apparatus implementing the video encoder performs the process 700.
- the video encoder receives (at step 710) a block of pixels of a video picture to be coded as a current block.
- the current block may be a CU, PU, or CTU of a video picture in a video sequence.
- the video encoder may receive the block of pixels from a video source.
- the video encoder performs (at step 720) an intra prediction operation for the current block according to an intra prediction setting and generate a set of prediction residuals (or residual pixels) .
- the intra prediction setting selects an intra prediction mode from a plurality of intra prediction modes as the current intra prediction mode for the current block.
- the intra prediction operation generates a set of prediction pixels by referencing pixels neighboring the current block in different directions according to the current intra prediction mode.
- the set of prediction residuals (or residual pixels) is generated based on the difference between the set of prediction pixels and the raw pixels of the current block.
- the intra prediction setting may be specified by a rate-distortion control operation.
- the intra prediction mode is selected from DC mode, planar mode, and multiple different directional/angular modes. Intra prediction modes are described in Section II above.
- the video encoder maps (at step 730) the current intra prediction mode to a target transform mode.
- the target transform mode may specify a horizontal transform type and a vertical transform type.
- a collectionof intra prediction modes are assigned with a particular transform mode, such that the target transform mode is the particular transform mode when the current intra prediction mode is in the collection of intra prediction modes.
- the collection of intra prediction modes may be defined by a fixed expression, such that different blocks with different block sizes have different fixed expression for defining the collection of intra prediction modes, or such that different fixed expressions are used to assign intra prediction modes in different numerical ranges with transform modes.
- intra prediction modes in a most probable mode (MPM) list of the current block are assigned with a particular transform mode.
- an intra prediction mode indicative of uniform block texture, such as DC mode is mapped to a particular transform mode.
- the MPM are described in Section III below. The mapping of intra prediction settings to transform settings is described in greater detail in Section IV below.
- the video encoder encodes (at step 740) the current block in a bitstream by performing a transform operation on the prediction residuals of the current block according to the target transform mode.
- Theprediction residuals of the current block are transformed into transform coefficients by the transform operation.
- the transform coefficients are entropy coded as syntax element of the bitstream.
- the target transform mode is not encoded in the bitstream as a syntax element.
- the intra prediction setting is also encoded in the bitstream as a syntax element.
- FIG. 8 illustrates an example video decoder 800 that may implicitly signal transform settings based on intra prediction settings.
- the video decoder 800 may implement the video decoder 200 of FIG. 2.
- the video decoder 800 is an image-decoding or video-decoding circuit that receives a bitstream 895 and decodes the content of the bitstream into pixel data of video frames for display.
- the video decoder 800 has several components or modules for decoding the bitstream 895, including an inverse quantization module 805, an inverse transform module 810, an intra prediction module 825, a motion compensation module 830, an in-loop filter 845, a decoded picture buffer 850, a MV buffer 865, a MV prediction module 875, and a parser 890.
- the motion compensation module 830 is part of an inter-prediction module 840.
- the modules 810 –890 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device. In some embodiments, the modules 810 –890 are modules of hardware circuits implemented by one or more ICs of an electronic apparatus. Though the modules 810 –890 are illustrated as being separate modules, some of the modules can be combined into a single module.
- the parser 890 receives the bitstream 895 and performs initial parsing according to the syntax defined by a video-coding or image-coding standard.
- the parsed syntax element includes various header elements, flags, as well as quantized data (or quantized coefficients) 812.
- the parser 890 parses out the various syntax elements by using entropy-coding techniques such as context-adaptive binary arithmetic coding (CABAC) or Huffman encoding.
- CABAC context-adaptive binary arithmetic coding
- Huffman encoding Huffman encoding
- the inverse quantization module 805 de-quantizes the quantized data (or quantized coefficients) 812 to obtain transform coefficients, and the inverse transform module 810 performs inverse transform on the transform coefficients 816 to produce reconstructed residual signal 819.
- the reconstructed residual signal 819 is added with predicted pixel data 813 from the intra prediction module 825 or the motion compensation module 830 to produce decoded pixel data 817.
- the decoded pixels data are filtered by the in-loop filter 845 and stored in the decoded picture buffer 850.
- the decoded picture buffer 850 is a storage external to the video decoder 800.
- the decoded picture buffer 850 is a storage internal to the video decoder 800.
- the intra prediction module 825 receives intra prediction data from bitstream 895 and according to which, produces the predicted pixel data 813 from the decoded pixel data 817 stored in the decoded picture buffer 850.
- the decoded pixel data 817 is also stored in a line buffer (not illustrated) for intra-picture prediction and spatial MV prediction.
- the content of the decoded picture buffer 850 is used for display.
- a display device 855 either retrieves the content of the decoded picture buffer 850 for display directly, or retrieves the content of the decoded picture buffer to a display buffer.
- the display device receives pixel values from the decoded picture buffer 850 through a pixel transport.
- the motion compensation module 830 produces predicted pixel data 813 from the decoded pixel data 817 stored in the decoded picture buffer 850 according to motion compensation MVs (MC MVs) . These motion compensation MVs are decoded by adding the residual motion data received from the bitstream 895 with predicted MVs received from the MV prediction module 875.
- MC MVs motion compensation MVs
- the MV prediction module 875 generates the predicted MVs based on reference MVs that were generated for decoding previous video frames, e.g., the motion compensation MVs that were used to perform motion compensation.
- the MV prediction module 875 retrieves the reference MVs of previous video frames from the MV buffer 865.
- the video decoder 800 stores the motion compensation MVs generated for decoding the current video frame in the MV buffer 865 as reference MVs for producing predicted MVs.
- the in-loop filter 845 performs filtering or smoothing operations on the decoded pixel data 817 to reduce the artifacts of coding, particularly at boundaries of pixel blocks.
- the filtering operation performed includes sample adaptive offset (SAO) .
- the filtering operations include adaptive loop filter (ALF) .
- FIG. 9 illustrates portions of the video decoder 800 that implement implicit derivation of transform settings based on intra prediction settings.
- intra prediction settings 900 control the operations of the intra prediction module 825.
- the intra prediction settings specify an intra prediction mode that is selected from DC mode, planar mode, and multiple different directional/angular modes.
- the intra prediction settings 900 may be implicit derived by the video decoder 800 (i.e., not encoded in the bitstream 895) .
- Intra prediction modes are described in Section II above.
- An implicit mapping module 910 receives the intra prediction settings 900 and maps the received intra prediction settings to transform settings for the inverse transform module 810.
- the transform settings may include a transform mode or transform types for a vertical transform and a horizontal transform.
- the transform types may be selected from various versions of DCT or DST such as DCT-II, DST-VII, DCT-VIII, DST-I, and DCT-V, or any subset of the above types.
- the inverse transform module 810 in turn performs transform on the prediction residuals of the block based on the derived transform settings.
- the intra prediction settings 900 are parsed out from the bitstream 895 by the entropy decoder 890 as syntax elements.
- the mapping of the transform settings from the intra prediction settings is described in Section IV above.
- FIG. 10 conceptually illustrates a video decoding process 1000 that implicitly decidestransform settings based on intra prediction settings.
- one or more processing units e.g., a processor
- a computing device implementing the video decoder 800 performs the process 1000 by executing instructions stored in a computer readable medium.
- an electronic apparatus implementing the video decoder performs the process 1000.
- the video decoder receives (at step 1010) a bitstream that includes coded prediction residuals (e.g., transform coefficients) of a current block of a video picture.
- the current block may be a CU, PU, or CTU of a video picture in a video sequence.
- the video decoder determines (at step 1020) a current intra prediction mode from a plurality of intra prediction modes for the current block according to an intra prediction setting.
- the intra prediction setting may be explicitly signaled bysyntax elements of the bitstream.
- the intra prediction mode is selected from DC mode, planar mode, and multiple different directional/angular modes. Alternatively, the intra prediction setting may be implicitly derived by the video decoder. Intra prediction modes are described in Section II above.
- the video decoder maps (at step 1030) the current intra prediction mode to a target transform mode.
- the target transform mode may specify a horizontal transform type and a vertical transform type.
- a collection of intra prediction modes are assigned with a particular transform mode, such that the target transform mode is the particular transform mode when the current intra prediction mode is in the collection of intra prediction modes.
- the collection of intra prediction modes may be defined by a fixed expression, such that different blocks with different block sizes have different fixed expression for defining the collection of intra prediction modes, or such that different fixed expressions are used to assign intra prediction modes in different numerical ranges with transform modes.
- intra prediction modes in a most probable mode (MPM) list of the current block are assigned with a particular transform mode.
- an intra prediction mode indicative of uniform block texture, such as DC mode is mapped to a particular transform mode.
- the MPM are described in Section III below. The mapping of intra prediction settings to transform settings is described in greater detail in Section IV below.
- the video decoder performs (at step 1040) an inverse transform operation on the coded prediction residuals of the current block according to the target transform mode to generate prediction residuals of the current block.
- the coded prediction residuals (or transform coefficients) are entropy decoded from syntax elements of the bitstream.
- the video decoder reconstructs (at step 1050) the current block based on the prediction residuals and a set of prediction pixels of the current block, e.g., by adding the set of residual pixels and the set of prediction pixels.
- an intra prediction operation generates the set of prediction pixels by referencing pixels neighboring the current block in different directions according to the current intra prediction mode.
- the current intra prediction mode is one of DC mode, planar mode, and multiple different directional/angular modes. Intra prediction modes are described in Section II above.
- Computer readable storage medium also referred to as computer readable medium
- these instructions are executed by one or more computational or processing unit (s) (e.g., one or more processors, cores of processors, or other processing units) , they cause the processing unit (s) to perform the actions indicated in the instructions.
- computational or processing unit e.g., one or more processors, cores of processors, or other processing units
- Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, random-access memory (RAM) chips, hard drives, erasable programmable read only memories (EPROMs) , electrically erasable programmable read-only memories (EEPROMs) , etc.
- the computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.
- the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage which can be read into memory for processing by a processor.
- multiple software inventions can be implemented as sub-parts of a larger program while remaining distinct software inventions.
- multiple software inventions can also be implemented as separate programs.
- any combination of separate programs that together implement a software invention described here is within the scope of the present disclosure.
- the software programs when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.
- FIG. 11 conceptually illustrates an electronic system 1100 with which some embodiments of the present disclosure are implemented.
- the electronic system 1100 may be a computer (e.g., a desktop computer, personal computer, tablet computer, etc. ) , phone, PDA, or any other sort of electronic device.
- Such an electronic system includes various types of computer readable media and interfaces for various other types of computer readable media.
- Electronic system 1100 includes a bus 1105, processing unit (s) 1110, a graphics-processing unit (GPU) 1115, a system memory 1120, a network 1125, a read-only memory 1130, a permanent storage device 1135, input devices 1140, and output devices 1145.
- the bus 1105 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the electronic system 1100.
- the bus 1105 communicatively connects the processing unit (s) 1110 with the GPU 1115, the read-only memory 1130, the system memory 1120, and the permanent storage device 1135.
- the processing unit (s) 1110 retrieves instructions to execute and data to process in order to execute the processes of the present disclosure.
- the processing unit (s) may be a single processor or a multi-core processor in different embodiments. Some instructions are passed to and executed by the GPU 1115.
- the GPU 1115 can offload various computations or complement the image processing provided by the processing unit (s) 1110.
- the read-only-memory (ROM) 1130 stores static data and instructions that are used by the processing unit (s) 1110 and other modules of the electronic system.
- the permanent storage device 1135 is a read-and-write memory device. This device is a non-volatile memory unit that stores instructions and data even when the electronic system 1100 is off. Some embodiments of the present disclosure use a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) as the permanent storage device 1135.
- the system memory 1120 is a read-and-write memory device. However, unlike storage device 1135, the system memory 1120 is a volatile read-and-write memory, such a random access memory.
- the system memory 1120 stores some of the instructions and data that the processor uses at runtime.
- processes in accordance with the present disclosure are stored in the system memory 1120, the permanent storage device 1135, and/or the read-only memory 1130.
- the various memory units include instructions for processing multimedia clips in accordance with some embodiments. From these various memory units, the processing unit (s) 1110 retrieves instructions to execute and data to process in order to execute the processes of some embodiments.
- the bus 1105 also connects to the input and output devices 1140 and 1145.
- the input devices 1140 enable the user to communicate information and select commands to the electronic system.
- the input devices 1140 include alphanumeric keyboards and pointing devices (also called “cursor control devices” ) , cameras (e.g., webcams) , microphones or similar devices for receiving voice commands, etc.
- the output devices 1145 display images generated by the electronic system or otherwise output data.
- the output devices 1145 include printers and display devices, such as cathode ray tubes (CRT) or liquid crystal displays (LCD) , as well as speakers or similar audio output devices. Some embodiments include devices such as a touchscreen that function as both input and output devices.
- CTR cathode ray tubes
- LCD liquid crystal displays
- bus 1105 also couples electronic system 1100 to a network 1125 through a network adapter (not shown) .
- the computer can be a part of a network of computers (such as a local area network ( “LAN” ) , a wide area network ( “WAN” ) , or an Intranet, or a network of networks, such as the Internet. Any or all components of electronic system 1100 may be used in conjunction with the present disclosure.
- Some embodiments include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media) .
- computer-readable media include RAM, ROM, read-only compact discs (CD-ROM) , recordable compact discs (CD-R) , rewritable compact discs (CD-RW) , read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM) , a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD-RW, DVD+RW, etc.
- the computer-readable media may store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.
- ASICs application specific integrated circuits
- FPGAs field programmable gate arrays
- integrated circuits execute instructions that are stored on the circuit itself.
- PLDs programmable logic devices
- ROM read only memory
- RAM random access memory
- the terms “computer” , “server” , “processor” , and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people.
- display or displaying means displaying on an electronic device.
- the terms “computer readable medium, ” “computer readable media, ” and “machine readable medium” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral signals.
- any two components so associated can also be viewed as being “operably connected” , or “operably coupled” , to each other to achieve the desired functionality, and any two components capable of being so associated can also be viewed as being “operably couplable” , to each other to achieve the desired functionality.
- operably couplable include but are not limited to physically mateable and/or physically interacting components and/or wirelessly interactable and/or wirelessly interacting components and/or logically interacting and/or logically interactable components.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Discrete Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP19785351.8A EP3777161A4 (fr) | 2018-04-13 | 2019-04-12 | Configuration de transformation implicite adaptative |
| BR112020020806A BR112020020806A8 (pt) | 2018-04-13 | 2019-04-12 | Configurações de transformação implícita adaptativa |
| CN201980016574.9A CN112042194B (zh) | 2018-04-13 | 2019-04-12 | 一种编码/解码的方法及电子装置 |
Applications Claiming Priority (6)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201862657100P | 2018-04-13 | 2018-04-13 | |
| US62/657,100 | 2018-04-13 | ||
| US201862665582P | 2018-05-02 | 2018-05-02 | |
| US62/665,582 | 2018-05-02 | ||
| US16/382,173 | 2019-04-11 | ||
| US16/382,173 US10999604B2 (en) | 2018-04-13 | 2019-04-11 | Adaptive implicit transform setting |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2019196941A1 true WO2019196941A1 (fr) | 2019-10-17 |
Family
ID=68160069
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2019/082537 Ceased WO2019196941A1 (fr) | 2018-04-13 | 2019-04-12 | Configuration de transformation implicite adaptative |
Country Status (6)
| Country | Link |
|---|---|
| US (1) | US10999604B2 (fr) |
| EP (1) | EP3777161A4 (fr) |
| CN (1) | CN112042194B (fr) |
| BR (1) | BR112020020806A8 (fr) |
| TW (1) | TWI692972B (fr) |
| WO (1) | WO2019196941A1 (fr) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2025167806A1 (fr) * | 2024-02-07 | 2025-08-14 | Beijing Dajia Internet Information Technology Co., Ltd. | Procédés et appareil de codage par transformée |
Families Citing this family (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2019072365A1 (fr) * | 2017-10-09 | 2019-04-18 | Huawei Technologies Co., Ltd. | Dispositifs et procédés de codage d'image et de vidéo |
| US10771778B2 (en) | 2018-09-14 | 2020-09-08 | Tencent America LLC | Method and device for MPM list generation for multi-line intra prediction |
| CN113365049B (zh) * | 2019-03-09 | 2023-02-28 | 杭州海康威视数字技术股份有限公司 | 进行编码和解码的方法、解码端、编码端和系统 |
| EP4397040A4 (fr) * | 2021-09-05 | 2025-07-23 | Beijing Dajia Internet Information Tech Co Ltd | Signalisation de modes de prédiction intra |
| JP2025010550A (ja) * | 2021-12-03 | 2025-01-22 | シャープ株式会社 | 動画像復号装置および動画像符号化装置 |
| WO2024212882A1 (fr) * | 2023-04-12 | 2024-10-17 | Mediatek Inc. | Procédés et appareil de transformée générale dans un mode de copie intra-bloc ou un mode d'appariement de modèle intra pour un codage vidéo |
| WO2025147924A1 (fr) * | 2024-01-10 | 2025-07-17 | Oppo广东移动通信有限公司 | Procédé de codage, procédé de décodage, flux de code, codeur, décodeur et support de stockage |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101069429A (zh) * | 2004-12-03 | 2007-11-07 | 三星电子株式会社 | 用于多层视频编码和解码的方法和设备 |
| WO2009113276A1 (fr) * | 2008-03-12 | 2009-09-17 | 三菱電機株式会社 | Dispositif de conversion pour système de codage d'image |
| US20120057630A1 (en) * | 2010-09-08 | 2012-03-08 | Samsung Electronics Co., Ltd. | Low complexity transform coding using adaptive dct/dst for intra-prediction |
| CN102857764A (zh) * | 2011-07-01 | 2013-01-02 | 华为技术有限公司 | 帧内预测模式处理的方法和装置 |
| CN103024391A (zh) * | 2010-01-07 | 2013-04-03 | 株式会社东芝 | 运动图像解码装置以及运动图像解码方法 |
| CN103931180A (zh) * | 2011-10-24 | 2014-07-16 | 码间私人有限公司 | 图像解码设备 |
| US20180020218A1 (en) | 2016-07-15 | 2018-01-18 | Qualcomm Incorporated | Look-up table for enhanced multiple transform |
Family Cites Families (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2010090749A1 (fr) * | 2009-02-06 | 2010-08-12 | Thomson Licensing | Procédés et appareil de signalisation de mode intra implicite et semi-implicite pour des codeurs et des décodeurs vidéo |
| US20110090952A1 (en) | 2009-10-21 | 2011-04-21 | Cohen Robert A | Directional Transforms for Video and Image Coding |
| KR101791242B1 (ko) | 2010-04-16 | 2017-10-30 | 에스케이텔레콤 주식회사 | 영상 부호화/복호화 장치 및 방법 |
| EP2590411A4 (fr) | 2010-07-02 | 2014-07-30 | Humax Co Ltd | Appareil et procédé destinés à coder/décoder des images pour un codage de prédiction intra |
| RU2719379C2 (ru) * | 2011-10-18 | 2020-04-17 | Кт Корпорейшен | Способ декодирования видеосигнала |
| ES2820297T3 (es) * | 2012-01-20 | 2021-04-20 | Dolby Laboratories Licensing Corp | Método de correlación de modo de intra predicción |
| US11477477B2 (en) | 2015-01-26 | 2022-10-18 | Qualcomm Incorporated | Sub-prediction unit based advanced temporal motion vector prediction |
| WO2018048265A1 (fr) | 2016-09-11 | 2018-03-15 | 엘지전자 주식회사 | Procédé et appareil de traitement de signal vidéo au moyen d'un vecteur de mouvement de flux optique amélioré |
| US10880564B2 (en) | 2016-10-01 | 2020-12-29 | Qualcomm Incorporated | Transform selection for video coding |
| US20190246134A1 (en) | 2018-02-06 | 2019-08-08 | Panasonic Intellectual Property Corporation Of America | Encoding method, decoding method, encoder, and decoder |
| US10567801B2 (en) * | 2018-03-07 | 2020-02-18 | Tencent America LLC | Method and apparatus for video coding with primary and secondary transforms |
| US10419754B1 (en) | 2018-04-02 | 2019-09-17 | Tencent America LLC | Method and apparatus for video decoding using multiple line intra prediction |
-
2019
- 2019-04-11 US US16/382,173 patent/US10999604B2/en active Active
- 2019-04-12 TW TW108112955A patent/TWI692972B/zh active
- 2019-04-12 CN CN201980016574.9A patent/CN112042194B/zh active Active
- 2019-04-12 BR BR112020020806A patent/BR112020020806A8/pt unknown
- 2019-04-12 WO PCT/CN2019/082537 patent/WO2019196941A1/fr not_active Ceased
- 2019-04-12 EP EP19785351.8A patent/EP3777161A4/fr not_active Withdrawn
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101069429A (zh) * | 2004-12-03 | 2007-11-07 | 三星电子株式会社 | 用于多层视频编码和解码的方法和设备 |
| WO2009113276A1 (fr) * | 2008-03-12 | 2009-09-17 | 三菱電機株式会社 | Dispositif de conversion pour système de codage d'image |
| CN103024391A (zh) * | 2010-01-07 | 2013-04-03 | 株式会社东芝 | 运动图像解码装置以及运动图像解码方法 |
| US20120057630A1 (en) * | 2010-09-08 | 2012-03-08 | Samsung Electronics Co., Ltd. | Low complexity transform coding using adaptive dct/dst for intra-prediction |
| CN102857764A (zh) * | 2011-07-01 | 2013-01-02 | 华为技术有限公司 | 帧内预测模式处理的方法和装置 |
| CN103931180A (zh) * | 2011-10-24 | 2014-07-16 | 码间私人有限公司 | 图像解码设备 |
| US20180020218A1 (en) | 2016-07-15 | 2018-01-18 | Qualcomm Incorporated | Look-up table for enhanced multiple transform |
Non-Patent Citations (2)
| Title |
|---|
| JIANLE CHEN ET AL.: "Algorithm description of Joint Exploration Test Model 7 (JEM7)", JVET OF ITU-T SG 16 WP 3 AND ISO/IEC JTC1/SC29/WG11, 13 June 2017 (2017-06-13) |
| See also references of EP3777161A4 |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2025167806A1 (fr) * | 2024-02-07 | 2025-08-14 | Beijing Dajia Internet Information Technology Co., Ltd. | Procédés et appareil de codage par transformée |
Also Published As
| Publication number | Publication date |
|---|---|
| CN112042194B (zh) | 2024-04-16 |
| TW201944779A (zh) | 2019-11-16 |
| US20190320204A1 (en) | 2019-10-17 |
| US10999604B2 (en) | 2021-05-04 |
| EP3777161A4 (fr) | 2021-04-28 |
| EP3777161A1 (fr) | 2021-02-17 |
| CN112042194A (zh) | 2020-12-04 |
| BR112020020806A2 (pt) | 2021-01-12 |
| BR112020020806A8 (pt) | 2023-02-07 |
| TWI692972B (zh) | 2020-05-01 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10887594B2 (en) | Entropy coding of coding units in image and video data | |
| US11297348B2 (en) | Implicit transform settings for coding a block of pixels | |
| US10999604B2 (en) | Adaptive implicit transform setting | |
| WO2018177300A1 (fr) | Prédiction de transformations multiples | |
| US11284077B2 (en) | Signaling of subpicture structures | |
| WO2021073521A1 (fr) | Signalisation de structure de partitionnement de blocs d'image et de vidéo | |
| US11228787B2 (en) | Signaling multiple transmission selection | |
| WO2023071778A1 (fr) | Modèle linéaire inter-composantes de signalisation | |
| WO2023198187A1 (fr) | Dérivation et prédiction de mode intra basées sur un modèle | |
| WO2023116704A1 (fr) | Prédiction de modèle linéaire trans-composante multi-modèle | |
| WO2023198105A1 (fr) | Dérivation et prédiction de mode intra implicites basées sur une région | |
| US11785214B2 (en) | Specifying video picture information | |
| WO2023236775A1 (fr) | Image de codage adaptative et données vidéo | |
| WO2023197998A1 (fr) | Types de partition de blocs étendus pour le codage vidéo | |
| WO2024022144A1 (fr) | Prédiction intra basée sur de multiples lignes de référence | |
| WO2023241340A1 (fr) | Matériel pour dérivation et prédiction de mode intra côté décodeur | |
| WO2023198110A1 (fr) | Partitionnement de blocs d'une image et de données vidéo | |
| WO2023241347A9 (fr) | Zones adaptatives pour dérivation et prédiction de mode intra côté décodeur | |
| WO2023208063A1 (fr) | Dérivation de modèle linéaire pour prédiction inter-composantes par de multiples lignes de référence | |
| WO2025016404A1 (fr) | Fusion de prédiction intra avec modèles de composantes transversales hérités | |
| WO2025148956A1 (fr) | Mélange basé sur une régression pour améliorer une prédiction intra avec un modèle voisin | |
| WO2023217235A1 (fr) | Affinement de prédiction avec modèle de convolution | |
| WO2025016275A1 (fr) | Stockage pour mode de fusion inter-composantes | |
| WO2021047590A1 (fr) | Signalisation de structures d'image secondaire | |
| CN121040060A (zh) | 用于视频和图像编解码的信令分割信息 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19785351 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112020020806 Country of ref document: BR |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2019785351 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 112020020806 Country of ref document: BR Kind code of ref document: A2 Effective date: 20201009 |