WO2025152827A1 - Procédés et appareil de dérivation de filtre pour prédiction intra basée sur un filtre dans un système de codage vidéo - Google Patents
Procédés et appareil de dérivation de filtre pour prédiction intra basée sur un filtre dans un système de codage vidéoInfo
- Publication number
- WO2025152827A1 WO2025152827A1 PCT/CN2025/071303 CN2025071303W WO2025152827A1 WO 2025152827 A1 WO2025152827 A1 WO 2025152827A1 CN 2025071303 W CN2025071303 W CN 2025071303W WO 2025152827 A1 WO2025152827 A1 WO 2025152827A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- filter
- block
- current block
- prediction
- current
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
Definitions
- VVC Versatile video coding
- JVET Joint Video Experts Team
- MPEG ISO/IEC Moving Picture Experts Group
- ISO/IEC 23090-3 2021
- Information technology -Coded representation of immersive media -Part 3 Versatile video coding, published Feb. 2021.
- VVC is developed based on its predecessor HEVC (High Efficiency Video Coding) by adding more coding tools to improve coding efficiency and also to handle various types of video sources including 3-dimensional (3D) video signals.
- HEVC High Efficiency Video Coding
- the transformed and quantized residues are then coded by Entropy Encoder 122 to be included in a video bitstream corresponding to the compressed video data.
- the bitstream associated with the transform coefficients is then packed with side information such as motion and coding modes associated with Intra prediction and Inter prediction, and other information such as parameters associated with loop filters applied to underlying image area.
- the side information associated with Intra Prediction 110, Inter prediction 112 and in-loop filter 130, is provided to Entropy Encoder 122 as shown in Fig. 1A. When an Inter-prediction mode is used, a reference picture or pictures have to be reconstructed at the encoder end as well.
- incoming video data undergoes a series of processing in the encoding system.
- the reconstructed video data from REC 128 may be subject to various impairments due to a series of processing.
- in-loop filter 130 is often applied to the reconstructed video data before the reconstructed video data are stored in the Reference Picture Buffer 134 in order to improve video quality.
- deblocking filter (DF) may be used.
- SAO Sample Adaptive Offset
- ALF Adaptive Loop Filter
- the loop filter information may need to be incorporated in the bitstream so that a decoder can properly recover the required information. Therefore, loop filter information is also provided to Entropy Encoder 122 for incorporation into the bitstream.
- DF deblocking filter
- SAO Sample Adaptive Offset
- ALF Adaptive Loop Filter
- Loop filter 130 is applied to the reconstructed video before the reconstructed samples are stored in the reference picture buffer 134.
- the system in Fig. 1A is intended to illustrate an exemplary structure of a typical video encoder. It may correspond to the High Efficiency Video Coding (HEVC) system, VP8, VP9, H. 264 or VVC.
- HEVC High Efficiency Video Coding
- the decoder can use similar or portion of the same functional blocks as the encoder except for Transform 118 and Quantization 120 since the decoder only needs Inverse Quantization 124 and Inverse Transform 126.
- the decoder uses an Entropy Decoder 140 to decode the video bitstream into quantized transform coefficients and needed coding information (e.g. ILPF information, Intra prediction information and Inter prediction information) .
- the Intra prediction 150 at the decoder side does not need to perform the mode search. Instead, the decoder only needs to generate Intra prediction according to Intra prediction information received from the Entropy Decoder 140.
- the decoder only needs to perform motion compensation (MC 152) according to Inter prediction information received from the Entropy Decoder 140 without the need for motion estimation.
- the decoder decodes the relevant syntax elements to determine the selected type of reconstructed area and the filter shape for the current block.
- the selected filter moves in the selected reconstructed area either horizontally or vertically with a one-pixel step to construct the auto-correlation matrix and the cross-correlation vector.
- the calculation of coefficients from the auto-correlation matrix and the cross-correlation vector is the same as that in convolutional cross-component model (CCCM) .
- CCCM convolutional cross-component model
- motion parameters consisting of motion vectors, reference picture indices and reference picture list usage index, and additional information needed for the new coding feature of VVC to be used for inter-predicted sample generation.
- the motion parameter can be signalled in an explicit or implicit manner.
- a CU is coded with skip mode, the CU is associated with one PU and has no significant residual coefficients, no coded motion vector delta or reference picture index.
- a merge mode is specified whereby the motion parameters for the current CU are obtained from neighbouring CUs, including spatial and temporal candidates, and additional schedules introduced in VVC.
- the merge mode can be applied to any inter-predicted CU, not only for skip mode.
- the alternative to merge mode is the explicit transmission of motion parameters, where motion vector, corresponding reference picture index for each reference picture list and reference picture list usage flag and other needed information are signalled explicitly per each CU.
- VTM 11 Versatile Video Coding and Test Model 11
- a method and apparatus for video coding are disclosed. According to this method, input data associated with a current block is received, wherein the input data comprise pixel data to be encoded at an encoder side or coded data associated with the current block to be decoded at a decoder side.
- a reference block of the current block is determined. Residuals of the reference block corresponding to differences between filter-predicted samples of the reference block and reconstructed samples of the reference block are determined, wherein the filter-predicted samples of the reference block are generated by applying a target filter-based intra prediction mode to the reference block.
- Target filter-based intra prediction associated with the target filter-based intra prediction mode is modified to target modified filter-based prediction for the current block according to the residuals of the reference block. Modified current filter-prediction is generated by applying the target modified filter-based prediction to the current block.
- the initial predictor is generated using intra prediction or non-intra prediction. In one embodiment, the initial predictor corresponds to the reconstructed samples or predicted samples prior to current to-be-filter-predicted samples. In one embodiment, the previous reconstructed block corresponds to one or more templates of one or more previous blocks. In one embodiment, the previous reconstructed block corresponds to an adjacent reference region or a non-adjacent region.
- Fig. 3 illustrates three types of filter shapes with fifteen inputs and generate one output for EIP process.
- Fig. 5 illustrates an example of scanning order for generating predictions for different positions in the current block by a diagonal order.
- Fig. 6 illustrates an example of the filter shape using a pattern around (not including) the position (x, y) of the to-be-predicted sample, where the pattern is a MxN region.
- Fig. 7 illustrates an example of a square MxN region excluding the region with the horizontal coordinate > x and excluding the region with the vertical coordinate > y.
- Fig. 8 illustrates an example of a non-square region (M ⁇ N) excluding the region with the horizontal coordinate > x and excluding the region with the vertical coordinate > y.
- Fig. 10 illustrates an example of template (i.e., neighbouring region) of the current block used for deriving filter parameters, where the template refers to top template, left template, and top-left template.
- Fig. 12 illustrates an example of to-be-predicted samples at the inner portion of the current block and/or far away from the top and left boundaries of the current block.
- Fig. 13 illustrates an example where the filter-prediction of the current block using the target filter-based intra prediction mode is combined with filter_pred_ref and the combined filter-prediction to generate the final prediction of the current block.
- Fig. 15 illustrates an example of using non-adjacent reference region for deriving filter parameters.
- Fig. 16 illustrates a flowchart of an exemplary video coding system that uses residuals of a reference block to modify the filter-prediction for the current block according to an embodiment of the present invention.
- a target filter-based intra prediction mode and/or a target filter-based novel prediction mode can be used to generate the prediction for an intra block or a non-intra block which refers to inter and/or block-vector mode.
- the block-vector mode refers to intra block copy or intra TMP, which searches in a pre-defined region of the current picture and finds a reference block (with the template of the reference block matching well with the template of the current block) in the pre-defined region to predict the current block.
- the filter using in the target-based intra prediction mode or the target-based novel prediction mode one way is from the derived path and/or another way is from the inheriting path.
- the predicted sample located at the position in the second, third, ..., n-th of the diagonal scanning for the current block is outputted in order, where n corresponds to the block size.
- the input of the filter is combined (or convolved) with the filter parameters to get the output of the filter.
- the input samples refer to reconstructed and/or filter-predicted and/or predicted samples associated with the current block and/or reconstructed and/or filter-predicted and/or predicted samples associated with the reference region reconstructed prior to or neighbouring to the current block.
- the filter-prediction of the current block which can be used to generate the final prediction of the current block, is obtained.
- the filter shape uses a pattern around (not including) the position (x, y) of the to-be-filter-predicted sample.
- the pattern is a MxN region or any subset of the MxN region around the to-be-filter-predicted sample where M is a first pre-defined positive number larger than 1 and N is a second pre-defined positive number larger than 1.
- Fig. 6 shows an example of the MxN region.
- M and N can be the same or different.
- M and N are equal to 1, 2, 3, 4, ..., or any number specified in the standard.
- M is larger than N if the current block has the block width larger than the block height.
- N is larger than M if the current block has the block height larger than the block width.
- M and/or N are adaptive according to the position (x, y) .
- the subset of the MxN region can be the MxN region (a) excluding the region with the horizontal coordinate > x and/or, (b) excluding the region with the vertical coordinate > y.
- the subset of MxN is the MxN region with (a) + (b) and M is equal to N. Therefore, the MxN region is square.
- the proposed method of subset of the MxN region is not limited to using in the case of M equal to N and/or can be applied to the case of M > N and/or N > M.
- the subset of MxN is the MxN region with (a) + (b) and M > N.
- the subset of MxN is the MxN region with (a) + (b) and N > M.
- FIG. 7 An example of the MxN region (square) excluding the region with the horizontal coordinate > x and excluding the region with the vertical coordinate > y is shown in Fig. 7.
- An example of the MxN region (non-square, M > N) excluding the region with the horizontal coordinate > x and excluding the region with the vertical coordinate > y is shown in Fig. 9.
- multiple filter shapes are supported and the selection from the multiple filter shapes depends on the explicit signalling at the block-level, CTU-level, slice-level, tile-level, SPS-level, PPS-level, picture-level (for example, picture header) , and/or sequence-level.
- a syntax is signalled to select a filter shape.
- the syntax is truncated unary coding.
- a shortest codeword is used to indicate the first candidate filter shape in the list.
- multiple determinations of filter parameters are supported and the selection from the multiple determinations of filter parameters depends on the explicit signalling at the block-level, CTU-level, slice-level, tile-level, SPS-level, PPS-level, picture-level (for example, picture header) , and/or sequence-level.
- the determination of filter parameters can be the derived path “deriving filter parameters using the template (neighbouring region) of the current block” and/or the inheriting path “inheriting filter information from the previously-coded blocks” .
- the selection between deriving filter parameters using the template (neighbouring region) of the current block and inheriting filter information from the previously-coded blocks depends on the explicit signalling at the block-level, CTU-level, slice-level, tile-level, SPS-level, PPS-level, picture-level (for example, picture header) , and/or sequence-level.
- the golden in response to deriving filter parameters using the template (neighbouring region) of the current block, the golden (i.e., a target for comparison) is the reconstructed samples on the template of the current block and the goal is to make the filtering result match the golden on the template.
- Any method to minimize the difference between the golden and the filter result can be used to get the filer parameters whose filtering result achieves a good matching with the golden.
- a regression-based method such as Gaussian elimination and/or any regression method unified with the methods of deriving cross-component models for the cross-component chroma modes can be used.
- chroma component samples can be predicted from corresponding reconstructed luma samples.
- the model parameters can be derived by minimizing the regression error between neighbouring luma and chroma samples.
- the template in response to deriving filter parameters using the template (neighbouring region) of the current block, the template refers to top template, left template, top-left template, and/or any combination of the above-mentioned templates as shown in Fig. 10.
- the selection of the template among multiple candidate templates can depend on the width, height, area, and/or position of the current block. For example, if the position of the current block is located at the top of picture boundary and/or the top of a CTU row boundary, the top template and/or the top-left template is not used for deriving filter parameters.
- the list size is adaptive according to how much filter information can be found using the all or any subset of the above-mention inherited candidates.
- the candidate with a smaller distortion on the template can be treated as a promising candidate during the measurement.
- the promising candidates are reordered to be put in the front of the list.
- One variation is that the syntax is not required to indicate a candidate from the reordered list. During the measurement, only the most promising candidate is kept and after checking each candidate in the list, the most promising candidate which has the smallest distortion is selected for the current block.
- Embodiment of the present invention as described above may be implemented in various hardware, software codes, or a combination of both.
- an embodiment of the present invention can be one or more circuit circuits integrated into a video compression chip or program code integrated into video compression software to perform the processing described herein.
- An embodiment of the present invention may also be program code to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
- DSP Digital Signal Processor
- the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA) .
- These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
- the software code or firmware code may be developed in different programming languages and different formats or styles.
- the software code may also be compiled for different target platforms.
- different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Des procédés et un appareil de codage vidéo utilisant un mode de prédiction intra basé sur un filtre sont divulgués. Selon ce procédé, un bloc de référence du bloc courant est déterminé. Des résidus du bloc de référence correspondant à des différences entre des échantillons prédits par filtre du bloc de référence et des échantillons reconstruits du bloc de référence sont déterminés, les échantillons prédits par filtre du bloc de référence étant générés par application d'un mode de prédiction intra basé sur un filtre cible au bloc de référence. La prédiction intra basée sur un filtre cible associée au mode de prédiction intra basé sur un filtre cible est modifiée en une prédiction basée sur un filtre modifiée cible pour le bloc courant en fonction des résidus du bloc de référence. Une prédiction basée sur un filtre courante modifiée est générée par application de la prédiction basée sur un filtre modifiée cible au bloc courant. Selon un autre procédé, des paramètres de filtre stockés calculé par une mise en correspondance de résultats filtrés avec des échantillons reconstruits d'un bloc reconstruit précédent sont utilisés pour générer une prédiction pour le bloc courant.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202463621640P | 2024-01-17 | 2024-01-17 | |
| US63/621,640 | 2024-01-17 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025152827A1 true WO2025152827A1 (fr) | 2025-07-24 |
Family
ID=96470699
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2025/071303 Pending WO2025152827A1 (fr) | 2024-01-17 | 2025-01-08 | Procédés et appareil de dérivation de filtre pour prédiction intra basée sur un filtre dans un système de codage vidéo |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2025152827A1 (fr) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109076237A (zh) * | 2015-11-18 | 2018-12-21 | 联发科技股份有限公司 | 在视频和图像压缩中使用帧内预测滤波器的帧内预测模式的方法和装置 |
| US20200275095A1 (en) * | 2019-02-27 | 2020-08-27 | Google Llc | Adaptive filter intra prediction modes in image/video compression |
| WO2022022622A1 (fr) * | 2020-07-29 | 2022-02-03 | Oppo广东移动通信有限公司 | Procédé de codage d'image, procédé de décodage d'image et appareil associé |
| US20220394273A1 (en) * | 2021-02-11 | 2022-12-08 | Qualcomm Incorporated | Upsampling reference pixels for intra-prediction in video coding |
-
2025
- 2025-01-08 WO PCT/CN2025/071303 patent/WO2025152827A1/fr active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109076237A (zh) * | 2015-11-18 | 2018-12-21 | 联发科技股份有限公司 | 在视频和图像压缩中使用帧内预测滤波器的帧内预测模式的方法和装置 |
| US20200275095A1 (en) * | 2019-02-27 | 2020-08-27 | Google Llc | Adaptive filter intra prediction modes in image/video compression |
| WO2022022622A1 (fr) * | 2020-07-29 | 2022-02-03 | Oppo广东移动通信有限公司 | Procédé de codage d'image, procédé de décodage d'image et appareil associé |
| US20220394273A1 (en) * | 2021-02-11 | 2022-12-08 | Qualcomm Incorporated | Upsampling reference pixels for intra-prediction in video coding |
Non-Patent Citations (1)
| Title |
|---|
| L. XU (OPPO), Y. YU (OPPO), H. YU (OPPO), J. GAN (OPPO), D. WANG (OPPO): "EE2-2.7: An extrapolation filter-based intra prediction mode", 32. JVET MEETING; 20231013 - 20231020; HANNOVER; (THE JOINT VIDEO EXPLORATION TEAM OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ), no. JVET-AF0080, 13 October 2023 (2023-10-13), XP030312107 * |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20220046239A1 (en) | Method and Apparatus of Encoding or Decoding Video Blocks with Constraints during Block Partitioning | |
| EP4222962A1 (fr) | Estimation de mouvement améliorée pour inter-codage | |
| WO2024017188A9 (fr) | Procédé et appareil pour un mélange de prédiction dans un système de codage vidéo | |
| WO2025021011A1 (fr) | Mode de prédiction combiné | |
| WO2024083115A1 (fr) | Procédé et appareil pour mélanger une prédiction intra et inter dans un système de codage vidéo | |
| WO2024083251A1 (fr) | Procédé et appareil de prédiction intra basée sur une zone à l'aide d'une dérivation de mode intra côté modèle ou décodeur dans un système de codage vidéo | |
| WO2023202713A1 (fr) | Procédé et appareil de dérivation de vecteur de mouvement en mode de fusion affine basée sur une régression dans de systèmes de codage vidéo | |
| WO2025152827A1 (fr) | Procédés et appareil de dérivation de filtre pour prédiction intra basée sur un filtre dans un système de codage vidéo | |
| WO2025148904A1 (fr) | Procédés et appareil de prédiction intra basée sur un filtre pour système de codage vidéo | |
| WO2025153050A1 (fr) | Procédés et appareil de prédiction intra basée sur un filtre avec de multiples hypothèses dans des systèmes de codage vidéo | |
| WO2025153010A1 (fr) | Prédiction basée sur un filtre | |
| WO2025082424A1 (fr) | Procédés et appareil de mode de fusion intra avec prédiction intra d'extrapolation | |
| WO2025218707A1 (fr) | Procédé et appareil de mode de fusion intra pour des modes mixtes à composantes de chrominance dans un système de codage vidéo | |
| WO2025082425A1 (fr) | Procédés et appareil de mode de prédiction combiné avec prédiction intra d'extrapolation pour codage vidéo | |
| WO2025077511A1 (fr) | Procédés et appareil d'héritage de modèle de compensation d'éclairage local pour codage vidéo | |
| WO2025223420A1 (fr) | Procédés et appareil de codage vidéo | |
| WO2025218691A1 (fr) | Procédés et appareil destinés à déterminer de manière adaptative un type de transformée sélectionné dans des systèmes de codage d'image et de vidéo | |
| WO2025223277A1 (fr) | Procédés et appareil de mode de fusion intra avec vérification de similarité dans un système de codage vidéo | |
| WO2025082073A1 (fr) | Procédés et appareil de dérivation et d'héritage de modèle de compensation d'éclairage local et non local pour codage vidéo | |
| WO2024193431A1 (fr) | Procédé et appareil de prédiction combinée dans un système de codage vidéo | |
| WO2024230472A1 (fr) | Procédés et appareil de fusion de mode intra dans une image et système de codage vidéo | |
| WO2024213093A1 (fr) | Procédés et appareil de mélange de prédiction intra pour codage vidéo | |
| WO2024193386A1 (fr) | Procédé et appareil de fusion de mode luma intra de modèle dans un système de codage vidéo | |
| WO2025007952A1 (fr) | Procédés et appareil d'amélioration de codage vidéo par dérivation de modèle | |
| WO2025237149A1 (fr) | Procédés et appareil de prédiction intra et de sélection de type de transformée dans des systèmes de codage d'image et de vidéo |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 25741389 Country of ref document: EP Kind code of ref document: A1 |