[go: up one dir, main page]

WO2013155664A1 - Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc) - Google Patents

Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc) Download PDF

Info

Publication number
WO2013155664A1
WO2013155664A1 PCT/CN2012/074124 CN2012074124W WO2013155664A1 WO 2013155664 A1 WO2013155664 A1 WO 2013155664A1 CN 2012074124 W CN2012074124 W CN 2012074124W WO 2013155664 A1 WO2013155664 A1 WO 2013155664A1
Authority
WO
WIPO (PCT)
Prior art keywords
depth levels
basis
leaf
parameter set
maximum
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2012/074124
Other languages
French (fr)
Inventor
Mei Guo
Shan Liu
Shaw-Min Lei
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MediaTek Singapore Pte Ltd
Original Assignee
MediaTek Singapore Pte Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by MediaTek Singapore Pte Ltd filed Critical MediaTek Singapore Pte Ltd
Priority to PCT/CN2012/074124 priority Critical patent/WO2013155664A1/en
Publication of WO2013155664A1 publication Critical patent/WO2013155664A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/463Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding

Definitions

  • the invention relates generally to video processing.
  • the present invention relates to methods and apparatuses of scalable extension of High Efficiency Video Coding (HEVC).
  • HEVC High Efficiency Video Coding
  • HEVC High Efficiency Video Coding
  • JCT-VC Joint Collaborative Team on Video Coding
  • HM-6.0 HEVC test model version 6.0
  • CU coding unit
  • Fig. 1 LCU with the maximum CU size 64x64 is recursively split into a series of Prediction Units (PUs) as shown in Fig.l.
  • PUs Prediction Units
  • a split flag for each CU indicating whether the current CU is further split is required to be transmitted.
  • Fig.2 gives the structure of PU within a LCU, which is produced with the recursive splitting in Fig.l.
  • PU can be further recursively split into a series of transform units (TU). Each TU is fed into transform separately. At each depth of TU splitting, a transform split flag for each TU indicating whether the current TU is further split is also required to be transmitted. When the TU size reaches the minimum one or the TU depth gets to the maximum one, the transform split flag is ignored.
  • JVT Joint Video Team
  • VCEG ITU-T Video coding Group
  • MPEG ISO/IEC Moving Picture Experts Group
  • the scalable extension allows multiple resolutions of a video sequence to be contained in a single bit stream (which is referred to as spatial scalability), together with multiple frame rates and multiple fidelity levels of a sequence within a single compressed bit stream (which are referred to as temporal scalability and quality scalability respectively).
  • spatial scalability multiple resolutions of a video sequence to be contained in a single bit stream
  • temporal scalability and quality scalability which are referred to as temporal scalability and quality scalability respectively.
  • the scalable extension of HEVC is being studied and is going to be developed under the JCT-VC group.
  • the scalable extension of H.264/AVC specifies a layered video coding system.
  • the motion-compensation and intra prediction are employed as for non- scalable coding.
  • the inter-layer prediction mechanisms are utilized in order to exploit the redundancy between layers.
  • a macroblock In a higher layer (with larger resolution or higher fidelity level), a macroblock, a residual signal and motion parameters, such as motion partitions, reference indices and motion vectors, could be predicted from the information of one lower layer (with smaller resolution or lower fidelity level).
  • the upsampling or the scaling is required in inter-layer prediction.
  • the scalable extension of H.264/AVC adopts a macroblock type denoted as Intra_BL.
  • Intra_BL a macroblock type denoted as Intra_BL.
  • the prediction signal of the macroblock in higher layer is derived from the corresponding reconstructed block in the reference layer. Upsampling or scaling is required to generate the prediction when two layers have different resolutions.
  • the CU structure of a lower layer can be reused as the collocated CU structure of a higher layer with or without signaling.
  • one flag is transmitted in the coding tree block or CU level of higher layer to indicate whether the current CU structure is derived from the corresponding CU in the reference layer or not. If yes, the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused.
  • the leaf CU of the CU structure that is derived from the reference layer can be further split into sub-CUs in higher layer.
  • Fig. 3 shows an example of reusing CU partitioning information, when the video resolution of higher layer is two times that of the reference lower layer in both spatial dimensions.
  • the CU structure of the corresponding CU in the reference layer is scaled by 2, and is used as the CU structure in higher layer. Moreover, the CU structure that is derived from the reference layer can be used as the initial coding structure in higher layer. In the leaf CU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf CU is divided into sub-CUs.
  • the TU structure of a lower layer can be reused as the collocated TU structure of a higher layer with or without signaling.
  • one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure is derived from the corresponding TU in the reference layer or not. If yes, the TU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused.
  • the TU structure that is derived from the reference layer can be used as the initial structure in higher layer.
  • one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs.
  • FIG. 1 is a figure illustrating the recursive CU partition in HM6.0
  • FIG. 2 is a figure illustrating the coding structure of CU which is generated with the recursive partition in Fig. 1;
  • Fig. 3 is a figure illustrating the CU structure of the reference lower layer is reused for the current higher layer.
  • the leaf CUs of the initial CU structure can be further split into sub-CUs in higher layer;
  • Fig. 4 is a figure illustrating the CU structure of the reference lower layer is reused for the current higher layer.
  • the leaf CUs of the initial CU structure can be further split into sub-CUs in higher layer with different additional depth levels.
  • Fig. 4 shows an example of further partitioning the leaf CUs in the initial CU structure into sub-CUs with several additional CU depth levels in higher layer.
  • the CUs "a" "b” and “c” are achieved with 1, 0, and 2 additional CU depth levels respectively.
  • the additional CU depth levels that are allowed in the current higher layer i.e., the maximum additional CU depth levels
  • the maximum additional CU depth levels corresponding to various CU depth levels or various sizes of the leaf CU could be stored in the codec. Besides, they can be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
  • the same number of the maximum additional CU depth levels is used for all leaf CUs in one LCU in higher layer. It is proposed to be determined based on the maximum or the average CU depth of all leaf CUs in one LCU.
  • the maximum additional CU depth levels corresponding to a diversity of maximum or average CU depth levels could be stored in the codec. Besides, they can be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
  • the additional CU depth levels that are allowed in the current higher layer i.e., the maximum additional CU depth levels
  • the maximum additional CU depth levels is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
  • the blocks in higher layer may be first predicted with the corresponding reconstructions in the reference lower layer. Upsampling or scaling is required to generate the prediction when two layers have different resolutions. Then the residual signal is further coded with the conventional intra or inter prediction in higher layer.
  • the CU structure of the reference lower layer can be reused as the collocated CU structure of the residual signal in higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the coding tree block or CU level of higher layer to indicate whether the current CU structure of the residual signal is derived from the corresponding CU in the reference layer or not.
  • the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused.
  • the leaf CU in the initial CU structure of the residual signal that is derived from the reference layer can be further split into sub-CUs.
  • one split flag can be transmitted to indicate whether the leaf CU is divided into sub-CUs.
  • the leaf CUs can be further split into sub-CUs with several additional CU depth levels.
  • the additional CU depth levels that are allowed in the current higher layer i.e., the maximum additional CU depth levels) is specified with the methods proposed in [0015].
  • the additional TU depth levels that are allowed in the current higher layer is proposed to be determined with the following three methods.
  • the additional TU depth levels that are allowed in the current higher layer i.e., the maximum additional TU depth levels
  • the maximum additional TU depth levels corresponding to various TU depth levels or various sizes of the leaf TU could be stored in the codec. Besides, they can also be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
  • the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU in higher layer. It is proposed to be determined based on the maximum or the average TU depth of all leaf TUs.
  • the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels or a diversity of average TU depth levels could be stored in the codec. Besides they can also be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header .
  • the number of the additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
  • the residual signal that is produced by the inter-layer intra prediction is recursively partitioned into a series of TUs before the transform and entropy coding.
  • the CU structure of a lower layer can be reused as the collocated TU structure of the residual signal in higher layer with or without signaling.
  • one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the residual signal is derived from the corresponding CU in the reference layer or not.
  • the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused.
  • the leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs.
  • one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs.
  • the leaf TUs can be further split into sub-TUs with several additional TU depth levels.
  • the additional TU depth levels that are allowed in the current higher layer i.e., the maximum additional TU depth levels
  • the residual signal that is produced by the inter-layer intra prediction is further coded with the conventional intra or inter prediction in higher layer.
  • the resulting residual signal is recursively partitioned into a series of TUs before the transform and entropy coding.
  • the CU or TU structure of a lower layer can be reused as the collocated TU structure of the residual signal in higher layer with or without signaling.
  • one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the residual signal is derived from the corresponding units (CU or TU) in the reference layer or not. If yes, the CU or TU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused.
  • the leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs.
  • one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs.
  • the leaf TUs can be further split into sub-TUs with several additional TU depth levels.
  • the additional TU depth levels that are allowed in the current higher layer i.e., the maximum additional TU depth levels
  • the methods described above can be used in a video encoder as well as in a video decoder.
  • Embodiments of the methods according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA).
  • These processors can be configured to perform particular tasks according to the invention, by executing machine- readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware codes may be developed in different programming languages and different format or style.
  • the software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
  • the invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Description

METHODS AND APPARATUSES OF CODING STRUCTURE FOR SCALABLE EXTENSION OF HIGH EFFICIENCY VIDEO
CODING (HEVC)
TECHNICAL FIELD
[0001] The invention relates generally to video processing. In particular, the present invention relates to methods and apparatuses of scalable extension of High Efficiency Video Coding (HEVC).
BACKGROUND
[0002] HEVC (High Efficiency Video Coding) is an advanced video coding system being developed under the Joint Collaborative Team on Video Coding (JCT-VC) group of video coding experts from ITU-T Study Group. In HEVC test model version 6.0 (HM-6.0), each frame is first partitioned into non- overlapped largest coding units (LCU) with the same size. The coding unit (CU) structure within a LCU is then expressed in a recursive tree representation, as exemplified in Fig. 1. LCU with the maximum CU size 64x64 is recursively split into a series of Prediction Units (PUs) as shown in Fig.l. At each depth except the maximum one, a split flag for each CU indicating whether the current CU is further split is required to be transmitted. Fig.2 gives the structure of PU within a LCU, which is produced with the recursive splitting in Fig.l.
[0003] In HM-6.0, PU can be further recursively split into a series of transform units (TU). Each TU is fed into transform separately. At each depth of TU splitting, a transform split flag for each TU indicating whether the current TU is further split is also required to be transmitted. When the TU size reaches the minimum one or the TU depth gets to the maximum one, the transform split flag is ignored.
[0004] The Joint Video Team (JVT), a joint organization of the ITU-T Video coding Group (VCEG) and the ISO/IEC Moving Picture Experts Group (MPEG), has developed a scalable extension to the H.264/AVC video coding standard, in order to support a diverse range of display devices and transmission channel capacities. The scalable extension allows multiple resolutions of a video sequence to be contained in a single bit stream (which is referred to as spatial scalability), together with multiple frame rates and multiple fidelity levels of a sequence within a single compressed bit stream (which are referred to as temporal scalability and quality scalability respectively). The scalable extension of HEVC is being studied and is going to be developed under the JCT-VC group.
[0005] The scalable extension of H.264/AVC specifies a layered video coding system. In each spatial layer or SNR layer, the motion-compensation and intra prediction are employed as for non- scalable coding. Moreover, the inter-layer prediction mechanisms are utilized in order to exploit the redundancy between layers. In a higher layer (with larger resolution or higher fidelity level), a macroblock, a residual signal and motion parameters, such as motion partitions, reference indices and motion vectors, could be predicted from the information of one lower layer (with smaller resolution or lower fidelity level). As for the spatial scalability, the upsampling or the scaling is required in inter-layer prediction.
[0006] The scalable extension of H.264/AVC adopts a macroblock type denoted as Intra_BL. When a macroblock in a higher layer is coded with Intra_BL and the collocated 8x8 submacroblock in its reference layer (i.e., a lower layer) is intra coded, the prediction signal of the macroblock in higher layer is derived from the corresponding reconstructed block in the reference layer. Upsampling or scaling is required to generate the prediction when two layers have different resolutions.
[0007] The CU structure of a lower layer can be reused as the collocated CU structure of a higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the coding tree block or CU level of higher layer to indicate whether the current CU structure is derived from the corresponding CU in the reference layer or not. If yes, the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused. The leaf CU of the CU structure that is derived from the reference layer can be further split into sub-CUs in higher layer. Fig. 3 shows an example of reusing CU partitioning information, when the video resolution of higher layer is two times that of the reference lower layer in both spatial dimensions. The CU structure of the corresponding CU in the reference layer is scaled by 2, and is used as the CU structure in higher layer. Moreover, the CU structure that is derived from the reference layer can be used as the initial coding structure in higher layer. In the leaf CU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf CU is divided into sub-CUs.
[0008] The TU structure of a lower layer can be reused as the collocated TU structure of a higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure is derived from the corresponding TU in the reference layer or not. If yes, the TU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused. Moreover, the TU structure that is derived from the reference layer can be used as the initial structure in higher layer. In the leaf TU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs.
SUMMARY
[0009] In light of the previously described problems, there exists a need for an apparatus and method, in which the coding unit structure, and the transform unit structure in higher layer can be determined based on the information from the reference lower layer. Some related information is required to be transmitted in parameter set or slice header.
BRIEF DESCRIPTION OF DRAWINGS
The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
[0010] Fig. 1 is a figure illustrating the recursive CU partition in HM6.0;
[0011] Fig. 2 is a figure illustrating the coding structure of CU which is generated with the recursive partition in Fig. 1;
[0012] Fig. 3 is a figure illustrating the CU structure of the reference lower layer is reused for the current higher layer. The leaf CUs of the initial CU structure can be further split into sub-CUs in higher layer;
[0013] Fig. 4 is a figure illustrating the CU structure of the reference lower layer is reused for the current higher layer. The leaf CUs of the initial CU structure can be further split into sub-CUs in higher layer with different additional depth levels.
DETAILED DESCRIPTION
[0014] The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
[0015] In the first embodiment, when the initial CU structure of a LCU in higher layer is derived from the reference lower layer, the number of the additional CU depth levels that are allowed in the current higher layer is proposed to be determined with the following three methods. Fig. 4 shows an example of further partitioning the leaf CUs in the initial CU structure into sub-CUs with several additional CU depth levels in higher layer. In Fig. 4, the CUs "a" "b" and "c" are achieved with 1, 0, and 2 additional CU depth levels respectively.
First, for each leaf CU, the additional CU depth levels that are allowed in the current higher layer (i.e., the maximum additional CU depth levels) is proposed to be determined based on the CU depth of the leaf CU in the initial CU structure or the size of leaf CU. The maximum additional CU depth levels corresponding to various CU depth levels or various sizes of the leaf CU could be stored in the codec. Besides, they can be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
Second, the same number of the maximum additional CU depth levels is used for all leaf CUs in one LCU in higher layer. It is proposed to be determined based on the maximum or the average CU depth of all leaf CUs in one LCU. The maximum additional CU depth levels corresponding to a diversity of maximum or average CU depth levels could be stored in the codec. Besides, they can be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
Third, the additional CU depth levels that are allowed in the current higher layer (i.e., the maximum additional CU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
[0016] In multi-loop scalable video coding system or the Intra_BL mode in single-loop scalable coding system, the blocks in higher layer may be first predicted with the corresponding reconstructions in the reference lower layer. Upsampling or scaling is required to generate the prediction when two layers have different resolutions. Then the residual signal is further coded with the conventional intra or inter prediction in higher layer. In the second embodiment, the CU structure of the reference lower layer can be reused as the collocated CU structure of the residual signal in higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the coding tree block or CU level of higher layer to indicate whether the current CU structure of the residual signal is derived from the corresponding CU in the reference layer or not. If yes, the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused. The leaf CU in the initial CU structure of the residual signal that is derived from the reference layer can be further split into sub-CUs. In the leaf CU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf CU is divided into sub-CUs. The leaf CUs can be further split into sub-CUs with several additional CU depth levels. The additional CU depth levels that are allowed in the current higher layer (i.e., the maximum additional CU depth levels) is specified with the methods proposed in [0015].
[0017] In the third embodiment, when the initial TU structure of a PU in higher layer is derived from the reference lower layer, the additional TU depth levels that are allowed in the current higher layer is proposed to be determined with the following three methods. First, for each leaf TU, the additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be determined based on the TU depth of the leaf TU in the initial TU structure or the size of leaf TU. The maximum additional TU depth levels corresponding to various TU depth levels or various sizes of the leaf TU could be stored in the codec. Besides, they can also be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
Second, the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU in higher layer. It is proposed to be determined based on the maximum or the average TU depth of all leaf TUs. The maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels or a diversity of average TU depth levels could be stored in the codec. Besides they can also be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header .
Third, the number of the additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
[0018] In multi-loop scalable video coding system or the Intra_BL mode in single-loop scalable coding system, the residual signal that is produced by the inter-layer intra prediction is recursively partitioned into a series of TUs before the transform and entropy coding. In the fourth embodiment, the CU structure of a lower layer can be reused as the collocated TU structure of the residual signal in higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the residual signal is derived from the corresponding CU in the reference layer or not. If yes, the CU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused. The leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs. In the leaf TU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs. The leaf TUs can be further split into sub-TUs with several additional TU depth levels. The additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is specified with the methods proposed in [0017].
[0019] In multi-loop scalable video coding system or the Intra_BL mode in single-loop scalable coding system, the residual signal that is produced by the inter-layer intra prediction is further coded with the conventional intra or inter prediction in higher layer. After the conventional intra or inter prediction, the resulting residual signal is recursively partitioned into a series of TUs before the transform and entropy coding.
In the fifth embodiment, the CU or TU structure of a lower layer can be reused as the collocated TU structure of the residual signal in higher layer with or without signaling. When it applies in explicit manner, one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the residual signal is derived from the corresponding units (CU or TU) in the reference layer or not. If yes, the CU or TU structure in reference lower layer is scaled (if the resolutions of two layers are different) and reused. The leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs. In the leaf TU of the initial structure of higher layer, one split flag can be transmitted to indicate whether the leaf TU is divided into sub-TUs. The leaf TUs can be further split into sub-TUs with several additional TU depth levels. The additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is specified with the methods proposed in [0017] [0020] The methods described above can be used in a video encoder as well as in a video decoder. Embodiments of the methods according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine- readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware codes may be developed in different programming languages and different format or style. The software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention. The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is intended to cover various
modifications and similar arrangements (as would be apparent to those skilled in the art). Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.

Claims

1. A method of Coding Unit (CU) partition in higher layer of scalable coding system, comprising:
determining additional CU depth levels that are allowed in a current higher layer, which is maximum additional CU depth levels, of leaf CUs in an initial CU structure that is derived from a corresponding CU in a reference lower layer.
2. The method as claimed in claim 1, wherein the maximum additional CU depth levels of each leaf CU are determined based on the CU depth of the leaf CU in the initial CU structure.
3. The method as claimed in claim 2, wherein the maximum additional CU depth levels corresponding to various CU depth levels are stored in the codec, and used for all video sequences.
4. The method as claimed in claim 2, wherein the maximum additional CU depth levels corresponding to various CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
5. The method as claimed in claim 1, wherein the maximum additional CU depth levels of each leaf CU are determined based on the size of the leaf CU in the initial CU structure.
6. The method as claimed in claim 5, wherein the maximum additional CU depth levels corresponding to various CU sizes are stored in the codec, and used for all video sequences.
7. The method as claimed in claim 5, wherein the maximum additional CU depth levels corresponding to various CU sizes are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
8. The method as claimed in claim 1, wherein the same number of the maximum additional CU depth levels is used for all leaf CUs in one Largest Coding Unit (LCU) of higher layer and it is determined based on the maximum CU depth of all leaf CUs in one LCU.
9. The method as claimed in claim 8, wherein the maximum additional CU depth levels corresponding to a diversity of maximum CU depth levels could be stored in the codec and used for all video sequences.
10. The method as claimed in claim 8, wherein the maximum additional CU depth levels corresponding to a diversity of maximum CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
11. The method as claimed in claim 1, wherein the same number of the maximum additional CU depth levels is used for all leaf CUs in one LCU of higher layer and it is determined based on the average CU depth levels of all leaf CUs in one LCU.
12. The method as claimed in claim 8, wherein the maximum additional CU depth levels corresponding to a diversity of average CU depth levels could be stored in the codec and used for all video sequences.
13. The method as claimed in claim 8, wherein the maximum additional CU depth levels corresponding to a diversity of average CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
14. The method as claimed in claim 1, wherein the additional CU depth levels that are allowed in the current higher layer (the maximum additional CU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
15. A method of CU partition in higher layer of scalable coding system in which the blocks in higher layer are first predicted with collocated reconstructions in a reference lower layer and then a residual signal is further coded with conventional intra or inter prediction in a higher layer, comprising: utilizing a collocated CU structure of the reference lower layer as a CU structure of the residual signal in the higher layer with or without signaling.
16. The method as claimed in claim 15, wherein one flag is transmitted in the coding tree block or CU level of higher layer to indicate whether the current CU structure of the residual signal is derived from the collocated CU in the reference layer or not.
17. The method as claimed in claim 15, wherein the CU structure in reference lower layer is scaled if the resolutions of two layers are different.
18. The method as claimed in claim 15, wherein the leaf CU in the initial CU structure of the residual signal that is derived from the reference layer can be further split into sub-CUs and one split flag is transmitted to indicate whether the leaf CU is divided into sub-CUs or not.
19. The method as claimed in claim 15, wherein the leaf CUs can be further split into sub-CUs with several additional CU depth levels.
20. The method as claimed in claim 19, wherein the maximum additional CU depth levels of each leaf CU in residual signal are determined based on the CU depth of the leaf CU in the initial CU structure.
21. The method as claimed in claim 20, wherein the maximum additional CU depth levels corresponding to various CU depth levels are stored in the codec, and used for all video sequences.
22. The method as claimed in claim 20, wherein the maximum additional CU depth levels corresponding to various CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
23. The method as claimed in claim 19, wherein the maximum additional CU depth levels of each leaf CU in residual signal are determined based on the size of the leaf CU in the initial CU structure.
24. The method as claimed in claim 23, wherein the maximum additional CU depth levels corresponding to various CU sizes are stored in the codec, and used for all video sequences.
25. The method as claimed in claim 23, wherein the maximum additional CU depth levels corresponding to various CU sizes are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
26. The method as claimed in claim 19, wherein the same number of the maximum additional CU depth levels is used for all leaf CUs in one LCU of residual signal in higher layer and it is determined based on the maximum CU depth levels of all leaf CUs in one LCU.
27. The method as claimed in claim 26, wherein the maximum additional CU depth levels corresponding to a diversity of maximum CU depth levels are stored in the codec and used for all video sequences.
28. The method as claimed in claim 26, wherein the maximum additional CU depth levels corresponding to a diversity of maximum CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
29. The method as claimed in claim 19, wherein the same number of the maximum additional CU depth levels is used for all leaf CUs in one LCU of residual signal in higher layer and it is determined based on the average CU depth of all leaf CUs in one LCU.
30. The method as claimed in claim 29, wherein the maximum additional CU depth levels corresponding to a diversity of average CU depth levels are stored in the codec and used for all video sequences.
31. The method as claimed in claim 29, wherein the maximum additional CU depth levels corresponding to a diversity of average CU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
32. The method as claimed in claim 19, wherein the additional CU depth levels of residual signal that are allowed in the current higher layer (i.e., the maximum additional CU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
33. A method of Transform Unit (TU) partition in a higher layer of scalable coding system, comprising:
determining additional TU depth levels that are allowed in the current higher layer, which is maximum additional TU depth levels, of leaf TUs in an initial TU structure of a Prediction Unit (PU) that is derived from a collocated PU in a reference lower layer.
34. The method as claimed in claim 33, wherein the maximum additional TU depth levels of each leaf TU are determined based on the TU depth of the leaf TU in the initial TU structure.
35. The method as claimed in claim 34, wherein the maximum additional TU depth levels corresponding to various TU depth levels are stored in the codec, and used for all video sequences.
36. The method as claimed in claim 34, wherein the maximum additional TU depth levels corresponding to various TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
37. The method as claimed in claim 33, wherein the maximum additional TU depth levels of each leaf TU are determined based on the size of the leaf TU in the initial TU structure.
38. The method as claimed in claim 37, wherein the maximum additional TU depth levels corresponding to various TU sizes are stored in the codec, and used for all video sequences.
39. The method as claimed in claim 37, wherein the maximum additional TU depth levels corresponding to various TU sizes are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
40. The method as claimed in claim 33, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU of higher layer and it is determined based on the maximum TU depth levels of all leaf TUs in one PU.
41. The method as claimed in claim 40, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels could be stored in the codec and used for all video sequences.
42. The method as claimed in claim 40, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
43. The method as claimed in claim 33, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU of higher layer and it is determined based on the average TU depth levels of all leaf TUs in one PU.
44. The method as claimed in claim 43, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels could be stored in the codec and used for all video sequences.
45. The method as claimed in claim 43, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
46. The method as claimed in claim 33, wherein the additional TU depth levels that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
47. A method of TU partition in higher layer of scalable coding system in which the CUs in higher layer are predicted with collocated reconstructions in a reference lower layer and then a residual signal is recursively partitioned into a series of TUs before the transform and entropy coding, comprising: utilizing a collocated CU structure of the reference lower layer as the TU structure of the residual signal in higher layer with or without signaling.
48. The method as claimed in claim 47, wherein one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the residual signal is derived from the collocated CU in the reference layer or not.
49. The method as claimed in claim 47, wherein the CU structure in reference lower layer is scaled if the resolutions of two layers are different.
50. The method as claimed in claim 47, wherein the leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs and one split flag is transmitted to indicate whether the leaf TU is divided into sub-TUs or not.
51. The method as claimed in claim 47, wherein the leaf TUs can be further split into sub-TUs with several additional TU depth levels.
52. The method as claimed in claim 51, wherein the maximum additional TU depth levels of each leaf TU in residual signal are determined based on the TU depth of the leaf TU in the initial TU structure.
53. The method as claimed in claim 52, wherein the maximum additional TU depth levels corresponding to various TU depth levels are stored in the codec, and used for all video sequences.
54. The method as claimed in claim 52, wherein the maximum additional TU depth levels corresponding to various TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
55. The method as claimed in claim 51, wherein the maximum additional TU depth levels of each leaf TU in residual signal are determined based on the size of the leaf TU in the initial TU structure.
56. The method as claimed in claim 55, wherein the maximum additional TU depth levels corresponding to various TU sizes are stored in the codec, and used for all video sequences.
57. The method as claimed in claim 55, wherein the maximum additional TU depth levels corresponding to various TU sizes are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
58. The method as claimed in claim 51, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one CU of residual signal in higher layer and it is determined based on the maximum TU depth of all leaf TUs in one CU.
59. The method as claimed in claim 58, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels are stored in the codec and used for all video sequences.
60. The method as claimed in claim 58, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
61. The method as claimed in claim 51, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one CU of residual signal in higher layer and it is determined based on the average TU depth levels of all leaf TUs in one CU.
62. The method as claimed in claim 61, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels are stored in the codec and used for all video sequences.
63. The method as claimed in claim 61, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
64. The method as claimed in claim 51, wherein the additional TU depth levels of residual signal that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
65. A method of TU partition in higher layer of scalable coding system in which the residual signal that is produced by inter-layer intra prediction is further coded with conventional intra or inter prediction in higher layer, and then PUs of a resulting residual signal is recursively partitioned into a series of TUs before transform and entropy coding, comprising: utilizing a collocated CU or TU structure of the reference lower layer as the TU structure of the residual signal in higher layer with or without signaling.
66. The method as claimed in claim 65, wherein one flag is transmitted in the root TU level or TU level of higher layer to indicate whether the current TU structure of the PU in the residual signal is derived from the collocated units (CU or TU) in the reference layer or not.
67. The method as claimed in claim 65, wherein the CU or TU structure in reference lower layer is scaled if the resolutions of two layers are different.
68. The method as claimed in claim 65, wherein the leaf TU in the initial TU structure of the residual signal that is derived from the reference layer can be further split into sub-TUs and one split flag is transmitted to indicate whether the leaf TU is divided into sub-TUs or not.
69. The method as claimed in claim 65, wherein the leaf TUs can be further split into sub-TUs with several additional TU depth levels.
70. The method as claimed in claim 69, wherein the maximum additional TU depth levels of each leaf TU in residual signal are determined based on the TU depth of the leaf TU in the initial TU structure.
71. The method as claimed in claim 70, wherein the maximum additional TU depth levels corresponding to various TU depth levels are stored in the codec, and used for all video sequences.
72. The method as claimed in claim 70, wherein the maximum additional TU depth levels corresponding to various TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
73. The method as claimed in claim 69, wherein the maximum additional TU depth levels of each leaf TU in residual signal are determined based on the size of the leaf TU in the initial TU structure.
74. The method as claimed in claim 73, wherein the maximum additional TU depth levels corresponding to various TU sizes are stored in the codec, and used for all video sequences.
75. The method as claimed in claim 73, wherein the maximum additional TU depth levels corresponding to various TU sizes are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
76. The method as claimed in claim 69, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU of residual signal in higher layer and it is determined based on the maximum TU depth levels of all leaf TUs in one PU.
77. The method as claimed in claim 76, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels are stored in the codec and used for all video sequences.
78. The method as claimed in claim 76, wherein the maximum additional TU depth levels corresponding to a diversity of maximum TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
79. The method as claimed in claim 69, wherein the same number of the maximum additional TU depth levels is used for all leaf TUs in one PU of residual signal in higher layer and it is determined based on the average TU depth of all leaf TUs in one PU.
80. The method as claimed in claim 79, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels are stored in the codec and used for all video sequences.
81. The method as claimed in claim 79, wherein the maximum additional TU depth levels corresponding to a diversity of average TU depth levels are adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
82. The method as claimed in claim 69, wherein the additional TU depth levels of residual signal that are allowed in the current higher layer (i.e., the maximum additional TU depth levels) is proposed to be adjusted on a sequence basis or a frame basis or a slice basis and signaled in sequence parameter set, picture parameter set or slice header.
PCT/CN2012/074124 2012-04-16 2012-04-16 Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc) Ceased WO2013155664A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2012/074124 WO2013155664A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2012/074124 WO2013155664A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc)

Publications (1)

Publication Number Publication Date
WO2013155664A1 true WO2013155664A1 (en) 2013-10-24

Family

ID=49382787

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2012/074124 Ceased WO2013155664A1 (en) 2012-04-16 2012-04-16 Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc)

Country Status (1)

Country Link
WO (1) WO2013155664A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104243997A (en) * 2014-09-05 2014-12-24 南京邮电大学 Method for quality scalable HEVC (high efficiency video coding)
CN111064968A (en) * 2019-12-18 2020-04-24 南华大学 VVC inter-frame CU deep rapid dividing method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011049397A2 (en) * 2009-10-23 2011-04-28 Samsung Electronics Co., Ltd. Method and apparatus for decoding video according to individual parsing or decoding in data unit level, and method and apparatus for encoding video for individual parsing or decoding in data unit level
WO2011071308A2 (en) * 2009-12-08 2011-06-16 Samsung Electronics Co., Ltd. Method and apparatus for encoding video by motion prediction using arbitrary partition, and method and apparatus for decoding video by motion prediction using arbitrary partition
WO2011087297A2 (en) * 2010-01-14 2011-07-21 Samsung Electronics Co., Ltd. Method and apparatus for encoding video by using deblocking filtering, and method and apparatus for decoding video by using deblocking filtering

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011049397A2 (en) * 2009-10-23 2011-04-28 Samsung Electronics Co., Ltd. Method and apparatus for decoding video according to individual parsing or decoding in data unit level, and method and apparatus for encoding video for individual parsing or decoding in data unit level
WO2011071308A2 (en) * 2009-12-08 2011-06-16 Samsung Electronics Co., Ltd. Method and apparatus for encoding video by motion prediction using arbitrary partition, and method and apparatus for decoding video by motion prediction using arbitrary partition
WO2011087297A2 (en) * 2010-01-14 2011-07-21 Samsung Electronics Co., Ltd. Method and apparatus for encoding video by using deblocking filtering, and method and apparatus for decoding video by using deblocking filtering

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104243997A (en) * 2014-09-05 2014-12-24 南京邮电大学 Method for quality scalable HEVC (high efficiency video coding)
CN104243997B (en) * 2014-09-05 2017-05-03 南京邮电大学 Method for quality scalable HEVC (high efficiency video coding)
CN111064968A (en) * 2019-12-18 2020-04-24 南华大学 VVC inter-frame CU deep rapid dividing method
CN111064968B (en) * 2019-12-18 2021-02-12 南华大学 VVC inter-frame CU deep rapid dividing method

Similar Documents

Publication Publication Date Title
US10791333B2 (en) Video encoding using hierarchical algorithms
KR102052290B1 (en) Deriving reference mode values and encoding and decoding information representing prediction modes
CN114097246B (en) Adaptive loop filtering across raster scan slices
EP2719183B1 (en) Method and apparatus of scalable video coding
CN103931196B (en) Inter-frame prediction method and device
KR102293097B1 (en) Devices and methods for video coding
US20140092956A1 (en) Adaptive transform options for scalable extension
KR20150137098A (en) Inter-layer reference picture restriction for high level syntax-only scalable video coding
WO2012167711A1 (en) Method and apparatus of scalable video coding
KR20230162801A (en) Externally enhanced prediction for video coding
KR102226893B1 (en) Video decoding method and apparatus using the same
EP2894854A2 (en) Image decoding method and apparatus using same
WO2013143103A1 (en) Methods and apparatuses of intra mode coding for scalable extension of high efficiency video coding (hevc)
WO2013155664A1 (en) Methods and apparatuses of coding structure for scalable extension of high efficiency video coding (hevc)
Hamidouche et al. Multi-core software architecture for the scalable HEVC decoder
EP4224857B1 (en) Deriving reference mode values and encoding and decoding information representing prediction modes
KR20160064843A (en) Method and apparatus for intra-coding in depth map
KR20130070215A (en) Method and apparatus for seletcing the adaptive depth information and processing deblocking filtering

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12874614

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12874614

Country of ref document: EP

Kind code of ref document: A1