[go: up one dir, main page]

WO2006004331A1 - Procedes de codage et de decodage video, codeur et decodeur video - Google Patents

Procedes de codage et de decodage video, codeur et decodeur video Download PDF

Info

Publication number
WO2006004331A1
WO2006004331A1 PCT/KR2005/002008 KR2005002008W WO2006004331A1 WO 2006004331 A1 WO2006004331 A1 WO 2006004331A1 KR 2005002008 W KR2005002008 W KR 2005002008W WO 2006004331 A1 WO2006004331 A1 WO 2006004331A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
intra
predicted
information
coding mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/KR2005/002008
Other languages
English (en)
Inventor
Sang-Chang Cha
Woo-Jin Han
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020040055283A external-priority patent/KR100654436B1/ko
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Publication of WO2006004331A1 publication Critical patent/WO2006004331A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • H04N19/615Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding using motion compensated temporal filtering [MCTF]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/19Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding using optimisation based on Lagrange multipliers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/53Multi-resolution motion estimation; Hierarchical motion estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • H04N19/64Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission
    • H04N19/647Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission using significance based coding, e.g. Embedded Zerotrees of Wavelets [EZW] or Set Partitioning in Hierarchical Trees [SPIHT]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]

Definitions

  • Apparatuses and methods consistent with the present invention relate to a video coding algorithm, and more particularly, to scalable video encoding and decoding capable of supporting an intra predictive coding mode.
  • a compression coding method is required for transmitting multimedia data including text, video, and audio.
  • a 24-bit true color image having a resolution of 640*480 needs a capacity of 640*480*24 bits, i.e., data of about 7.37 Mbits, per frame.
  • a bandwidth of 221 Mbits/sec is required.
  • a 90- minute movie based on such an image is stored, a storage space of about 1200 Gbits is required.
  • a compression coding method is a requisite for transmitting multimedia data including text, video, and audio.
  • Data redundancy is typically defined as: (i) spatial redundancy in which the same color or object is repeated in an image; (ii) temporal redundancy in which there is little change between adjacent frames in a moving image or the same sound is repeated in audio; or (iii) mental visual redundancy taking into account human eyesight and perception dull to high frequency. Data can be compressed by removing such data redundancy.
  • Data compression can largely be classified into lossy/lossless compression, according to whether source data is lost, intraframe/interframe compression, according to whether individual frames are compressed independently, and symmetric/asymmetric compression, according to whether a time required for compression is the same as a time required for recovery.
  • data compression is defined as real-time compression when a compression/ recovery time delay does not exceed 50 ms and as scalable compression when frames have different resolutions.
  • lossless compression is usually used for text or medical data.
  • lossy compression is usually used for multimedia data.
  • intraframe compression is usually used to remove spatial redundancy
  • interframe compression is usually used to remove temporal redundancy.
  • Transmission performance is different depending on transmission media.
  • Currently used transmission media have various transmission rates. For example, an ultra high ⁇ speed communication network can transmit data of several tens of megabits per second while a mobile communication network has a transmission rate of 384 kilobits per second.
  • video coding methods such as Motion Picture Experts Group (MPEG)-I, MPEG-2, H.263, and H.264, temporal redundancy is removed by motion compensation based on motion estimation and compensation, and spatial redundancy is removed by transform coding. These methods have satisfactory compression rates, but they do not have the flexibility of a truly scalable bitstream since they use a reflexive approach in a main algorithm.
  • MPEG Motion Picture Experts Group
  • Scalability indicates the ability to partially decode a single compressed bitstream, that is, the ability to perform a variety of types of video reproduction.
  • Scalability includes spatial scalability indicating a video resolution, signal-to noise ratio (SNR) scalability indicating a video quality level, temporal scalability indicating a frame rate, and a combination thereof.
  • SNR signal-to noise ratio
  • MCTF motion compensated temporal filtering
  • FIG. 1 is a block diagram of an MCTF-based scalable video encoder
  • FIG. 2 il ⁇ lustrates a temporal filtering process in conventional MCTF-based video coding.
  • a scalable video encoder includes a motion estimator 110 estimating motion between input video frames and determining motion vectors, a motion compensated temporal filter 140 compensating the motion of an interframe using the motion vectors and removing temporal redundancies within the interframe subjected to motion compensation, a spatial transformer 150 removing spatial re ⁇ dundancies within an intraframe and the interframe within which the temporal re ⁇ dundancies have been removed and producing transform coefficients, a quantizer 160 quantizing the transform coefficients in order to reduce the amount of data, a motion vector encoder 120 encoding a motion vector in order to reduce bits required for the motion vector, and a bitstream generator 130 using the quantized transform co ⁇ efficients and the encoded motion vectors to generate a bitstream.
  • the motion estimator 110 calculates a motion vector to be used in compensating the motion of a current frame and removing temporal redundancies within the current frame.
  • the motion vector is defined as a displacement from the best-matching block in a reference frame with respect to a block in a current frame.
  • HVSBM Hierarchical Variable Size Block Matching
  • a frame having an N*N resolution is first downsampled to form frames with lower resolutions such as N/2*N/2 and N/4*N/4 resolutions. Then, a motion vector is obtained at the N/4*N/4 resolution and a motion vector having N/2*N/2 resolution is obtained using the N/4*N/4 resolution motion vector. Similarly, a motion vector with N*N resolution is obtained using the N/2*N/2 resolution motion vector.
  • the final block size and the final motion vector are determined through a selection process.
  • the motion compensated temporal filter 140 removes temporal redundancies within a current frame using the motion vectors obtained by the motion estimator 110. To accomplish this, the motion compensated temporal filter 140 uses a reference frame and motion vectors to generate a predicted frame and compares the current frame with the predicted frame to thereby generate a residual frame. The temporal filtering process will be described in more detail later with reference to FIG. 2.
  • the spatial transformer 150 spatially transforms the residual frames to obtain transform coefficients.
  • the video encoder removes spatial redundancies within the residual frames using wavelet transform.
  • the wavelet transform is used to generate a spatially scalable bitstream.
  • the quantizer 160 uses an embedded quantization algorithm to quantize the transform coefficients obtained through the spatial transformer 150.
  • Embedded quantization algorithms currently known are Embedded Zerotree Wavelet (EZW), Set Partitioning in Hierarchical Trees (SPIHT), Embedded Zero Block Coding (EZBC), and Embedded Block Coding with Optimized Truncation (EBCOT).
  • EZW Embedded Zerotree Wavelet
  • SPIHT Set Partitioning in Hierarchical Trees
  • EZBC Embedded Zero Block Coding
  • EBCOT Embedded Block Coding with Optimized Truncation
  • any one among the known embedded quantization algorithms may be used.
  • Embedded quantization is used to generate bitstreams having SNR scalability.
  • the motion vector encoder 120 encodes the motion vectors calculated by the motion estimator 110.
  • the bitstream generator 130 generates a bitstream containing the quantized transform coefficients and the encoded motion vectors.
  • a group of picture (GOP) size is assumed to be 16.
  • a scalable video encoder receives 16 frames and performs MCTF forward with respect to the 16 frames, thereby obtaining 8 low-pass frames and 8 high-pass frames. Then, in temporal level 1, MCTF is performed forward with respect to the 8 low-pass frames, thereby obtaining 4 low-pass frames and 4 high-pass frames. In temporal level 2, MCTF is performed forward with respect to the 4 low-pass frames obtained in temporal level 1, thereby obtaining 2 low-pass frames and 2 high- pass frames. Lastly, in temporal level 3, MCTF is performed forward with respect to the 2 low-pass frames obtained in temporal level 2, thereby obtaining 1 low-pass frame and 1 high-pass frame.
  • the video encoder predicts motion between the two frames, generates a predicted frame by compensating the motion, compares the predicted frame with one frame to thereby generate a high- pass frame, and calculates the average of the predicted frame and the other frame to thereby generate a low-pass frame.
  • the decoder decodes the frame LLLL 16 to reconstruct a video sequence with a frame rate that is one sixteenth of the frame rate of the original video sequence.
  • the decoder decodes the frames LLLL 16 and LLLH8 to re ⁇ construct a video sequence with a frame rate that is one eighth of the frame rate of the original video sequence.
  • the decoder reconstructs video sequences with a quarter frame rate, a half frame rate, and a full frame rate from a single bitstream.
  • scalable video coding allows the decoder to generate video sequences at various resolutions, various frames rates or various qualities from a single bitstream, this technique can be used in a wide variety of applications.
  • currently known scalable video coding schemes offer significantly lower compression efficiency than other existing coding schemes such as H.264. Since the low compression efficiency is an important factor that severely impedes the wide use of scalable video coding, various attempts are being made to improve compression efficiency for scalable video coding.
  • One of the various approaches is to introduce an intra predictive coding mode into an MCTF process.
  • the present invention provides scalable video encoding and decoding methods capable of supporting an intra predictive coding mode and a scalable video encoder and a scalable video decoder.
  • a video encoding method including: determining one of inter predictive coding and intra predictive coding modes as a coding mode for each block in an input video frame; generating a predicted frame for the input video frame using predicted blocks obtained according to the determined coding mode; and encoding the input video frame using the predicted frame.
  • the intra predictive coding mode is determined as the coding mode, an intra basis block composed of representative values of a block is generated for a block and the intra basis block is interpolated to generate an intra predicted block for the block.
  • a video encoder including a mode determiner determining one of an inter predictive coding mode and an intra predictive coding mode as a coding mode for each block in an input video frame and generating predicted blocks according to the determined mode, a temporal filter generating a predicted frame for the input video frame using the predicted blocks and removing temporal redundancies within the video frame using the predicted frame, a spatial transformer removing spatial redundancies within the video frame in which the temporal redundancies have been removed, a quantizer quantizing the video frame in which the spatial redundancies have been removed, and a bitstream generator generating a bitstream containing the quantized video frame, wherein the mode determiner generates an intra basis block composed of representative values for a block for which an intra predictive coding mode is determined and then generates an intra predicted block for the block by interpolating the intra basis block.
  • a video decoding method including interpreting an input bitstream and obtaining texture in ⁇ formation, motion vector information, and intra basis block information, generating a predicted frame using the texture information, the motion vector information, and the intra basis block information, and reconstructing a video frame using the predicted frame, wherein an intra predicted block in the predicted frame is obtained by adding residual block information contained in the texture information to intra predicted block information obtained by interpolating the intra basis block information.
  • a video decoder including a bitstream interpreter interpreting a bitstream and obtaining texture information, motion vector information, and intra basis block information, an inverse quantizer inversely quantizing the texture information, an inverse spatial transformer performing inverse spatial transform on the inversely quantized texture information and generating a residual frame, and an inverse temporal filter generating a predicted frame using the residual frame, the motion vector information, and the intra basis block information and reconstructing a video frame using the predicted frame, wherein the inverse temporal filter generates an intra predicted block in the predicted frame by adding residual block information contained in the residual frame to intra predicted block information obtained by interpolating the intra basis block information.
  • FIG. 1 is a block diagram of a conventional scalable video encoder
  • FIG. 2 illustrates a temporal filtering process in conventional scalable video coding
  • FIG. 3 is a block diagram of a video encoder according to an exemplary embodiment of the present invention.
  • FIG. 4 is a diagram for explaining a process of generating an intra basis block according to an exemplary embodiment of the present invention.
  • FIG. 5 is a diagram for explaining a process of generating an intra predicted block according to an exemplary embodiment of the present invention
  • FIG. 6 is a diagram for explaining a process of filtering a predicted frame according to an exemplary embodiment of the present invention.
  • FIG. 7 illustrates the process of an intra predictive coding mode according to an exemplary embodiment of the present invention
  • FIG. 8 illustrates the process of an intra predictive coding mode according another exemplary embodiment of the present invention.
  • FIG. 9 is a block diagram of a video decoder according to an exemplary embodiment of the present invention.
  • Video coding algorithms employ intra prediction and frame filtering techniques to improve coding efficiency and image quality, respectively.
  • Intra prediction can be used for scalable video coding algorithms as well as discrete cosine transform (DCT)-based video coding algorithms.
  • the intra prediction and the frame filtering can be performed inde ⁇ pendently or together.
  • the present invention will be described with reference to exemplary embodiments in which scalable video coding uses intra- prediction and frame filtering together.
  • some components may be optional or can be replaced by other components performing different functions.
  • FIG. 3 is a block diagram of a video encoder supporting an intra predictive coding mode according to an exemplary embodiment of the present invention.
  • the video encoder includes a mode determiner 310, a temporal filter 320, a wavelet transformer 330, a quantizer 340, and a bitstream generator 350.
  • the mode determiner 310 determines a mode in which each block in a frame currently being encoded ('current frame') will be encoded. To accomplish this function, the mode determiner 310 includes an inter prediction unit 312, an intra prediction unit 314, and a determination unit 316.
  • the inter prediction unit 312 estimates motion between each block in the current frame and a corresponding reference block using one or more reference frames and obtains a motion vector. Following the motion estimation, the inter prediction unit 312 calculates a difference metric between the block and the corresponding reference block. While a mean of absolute difference (MAD) is used as the difference metric in the present invention, sum of absolute difference (SAD) or other metrics may be used.
  • the difference metric is used to calculate a cost for a coding scheme.
  • the intra prediction unit 314 encodes each block in the current frame using in ⁇ formation within the current frame.
  • An intra predictive coding mode is used in the present exemplary embodiment to generate an intra predicted block for each block in the current frame with reference to an intra basis block for the block and calculate a difference metric between the block and the corresponding intra predicted block.
  • a process of generating an intra basis block and an intra predicted block will be described in more detail later.
  • the determination unit 316 receives difference metrics for each block in the current frame from the inter prediction unit 312 and the intra prediction unit 314 and determines a coding mode for the block. For example, to determine the coding mode for each block, the determination unit 316 may compare costs for an intra predictive coding mode and an inter predictive mode. Costs C and C for inter predictive inter intra coding and intra predictive coding a block are defined by Equation (1) as follows:
  • C intra D mtra + ⁇ (INTRA_bits+Mode_bits intra ) ... (1)
  • D is a difference metric between the block and a corresponding reference block inter for inter predictive coding and D is a difference metric between the block and a cor- mtra responding intra predicted block for intra-coding.
  • MV_bits and INTRA_bits re ⁇ spectively denote the number of bits allocated to a motion vector associated with the block and the intra basis block.
  • Mode_bits and Mode_bits denote the number of inter intra bits required to indicate that the block is encoded as an inter-block and intra-block, re ⁇ spectively, ⁇ is a Lagrangian coefficient used to control the balance among the bits allocated to a motion vector and a texture (image).
  • the determination unit 316 can determine the mode in which each block in the current frame will be encoded. For example, when a cost for inter predictive coding is less than a cost for intra predictive coding, the determination unit 316 determines that the block will be inter-coded. Conversely, when the cost for intra predictive coding is less than the cost for inter predictive coding, the de ⁇ termination unit 316 determines that the block will be intra-coded.
  • the temporal filter 320 generates a predicted frame for the current frame, compares the current frame with the predicted frame, and removes temporal redundancies within the current frame.
  • the temporal filter 320 may also remove block artifacts that can be generated during prediction (inter prediction or intra prediction). The block artifacts that appear along block boundaries in the predicted frame generated on a block-by-block basis sig ⁇ nificantly degrade the visual quality of image.
  • the temporal filter 320 includes a predicted frame filtering unit 324 removing block artifacts in the predicted frame.
  • the predicted frame filtering unit 324 may perform filtering on the predicted frame to remove a block artifact introduced at a boundary between an intra predicted block and an inter predicted block as well as a block artifact at a boundary between inter predicted blocks.
  • the predicted frame filtering unit 324 can be used for a video coding algorithm not supporting an intra predictive coding mode.
  • the temporal filter 320 may further include an updating unit 326 when scalable video coding includes the operation of updating frames.
  • the updating unit 326 is not required for scalable video coding which does not include the updating operation or DCT-based video coding.
  • the predicted frame generating unit 322 generates a predicted frame using a reference block or an intra-predicted block corresponding to each block in a current frame.
  • a comparator compares the current frame with the predicted frame to thereby generate a residual frame.
  • the predicted frame filtering unit 324 performs filtering on the predicted frame to reduce block artifacts that can occur in the residual frame. That is, the comparator compares the current frame with the predicted frame subjected to filtering, thereby generating the residual frame.
  • a process of filtering the predicted frame will be described in more detail later.
  • a filtering process for the predicted frame was mostly used for closed-loop video coding such as H.264 video coding schemes. The filtering process was not used for open-loop scalable video coding that allows an encoded bitstream to be truncated by a predecoder for decoding.
  • the open-loop scalable video coding did not employ filtering of a predicted frame.
  • scalable video coding including filtering of a predicted frame provides improved video quality. Therefore, the present invention includes the operation of filtering a predicted frame.
  • the updating unit 326 updates the residual frames (H frames) and original video frames in an MCTF-based scalable video coding algorithm and generates a single low- pass subband (L frame) and a plurality of high-pass subbands (H frames).
  • L frame low- pass subband
  • H frames high-pass subbands
  • FIG. 2 residual frames obtained from frames 1, 3, 5, 7, 9, 11, 13, and 15, and frames 2, 4, 6, 8, 10, 12, 14, and 16 are updated to generate subbands in temporal level 1.
  • L frames in temporal level 1 are subjected to motion estimation or intra prediction by the mode determiner 310, pass through the predicted frame generating unit 322 and the predicted frame filtering unit 324, and are input into the updating unit 326.
  • the updating unit 326 generates subbands (L frames and H frames) in temporal level 2 using residual frames from the L frames in temporal level 1 and the L frames in temporal level 1.
  • the L frames in temporal level 2 is used to generate subbands in temporal level 3.
  • L frames in temporal level 3 is used to a single H frame and a single L frame in temporal level 4. While the updating operation is performed by a 5/3 filter, a Haar filter or a 7/5 filter may be used as is conventionally done.
  • the wavelet transformer 330 performs wavelet transform on the frames subjected to temporal filtering by the temporal filter 320.
  • a frame is decomposed into four sections (quadrants).
  • a quarter-sized image (L image) which is substantially the same as the entire image, appears in a quadrant of the frame, and information (H image), which is needed to reconstruct the entire image from the L image, appears in the other three quadrants.
  • the L image may be decomposed into a quarter-sized LL image and information needed to reconst ruct the L image.
  • Image compression based on the wavelet transform is applied to JPEG 2000 compression technique. Spatial redundancy of a frame can be removed by wavelet transform.
  • wavelet transform unlike in the DCT transform, original image data is stored in a size-reduced form.
  • the sized-reduced image enables spatially scalable video coding. While it is described above in the exemplary embodiment illustrated in FIG. 3 that wavelet transform is used as a spatial trans ⁇ formation technique in scalable video coding supporting an intra predictive coding mode, DCT may also be used when the intra predictive coding mode is applied to the existing video coding standards such as MPEG-2, MPEG-4, and H.264.
  • the quantizer 340 uses an embedded quantization algorithm to quantize the wavelet transformed frames.
  • the embedded quantization involves quantization, scanning, and entropy coding. Texture information that will be contained in a bitstream is generated by the embedded quantization.
  • a motion vector that should be also contained in the bitstream in order to decode a block encoded in an inter predictive mode may be encoded using lossless compression.
  • a motion vector encoder 360 encodes a motion vector obtained from the inter prediction unit 314 using variable length coding or arithmetic coding and transmits the encoded motion vector to the bitstream generator 350.
  • the bitstream also contains an intra basis block in order to decode a block encoded in an intra predictive coding mode.
  • the intra basis block Before being transmitted to the bitstream generator 350, the intra basis block is not compressed or encoded. Alternatively, the intra basis block may be quantized or be encoded using variable length coding or arithmetic coding.
  • the video encoder of FIG. 3 uses a quantized intra basis block. More specifically, when a block is encoded in an intra predictive coding mode, the intra prediction unit 314 generates an intra basis block for the block and an intra predicted block using the intra basis block.
  • the intra prediction unit 314 obtains a difference metric by comparing the block with the intra predicted block and transmits the difference metric to the determination unit 316.
  • the determination unit 316 determines that the block is encoded in an intra predictive coding mode, the intra predicted block is provided to the temporal filter 420.
  • the intra prediction unit 314 predicts an intra basis block from neighboring subblocks surrounding the block and generates a residual intra basis block by comparing the predicted intra basis block with the original intra basis block.
  • the intra quantization unit 370 quantizes the residual intra basis block in order to reduce the amount of information and sends the quantized residual intra basis block back to the intra prediction unit 314.
  • the quantization may include a trans ⁇ formation operation to reduce the amount of information in the residual intra basis block.
  • the intra prediction unit 314 adds the quantized residual intra basis block to the intra basis block predicted from the neighboring subblocks and generates a new intra basis block.
  • the intra prediction unit 314 then generates an intra predicted block by in ⁇ terpolating the new intra basis block and transmits the intra predicted block to the temporal filter 320 in order to be used in generating residual blocks.
  • the temporal filter 320 After generating a predicted frame using intra predicted blocks and inter predicted blocks, the temporal filter 320 compares the predicted frame with an original video frame to thereby generate a residual frame.
  • the residual frame passes through the wavelet transformer 330 and the quantizer 340 and is combined into a bitstream.
  • the bitstream generator 350 generates a bitstream using texture information received from the quantizer 340, motion vectors received from the motion vector encoder 360, and quantized intra basis blocks received from the intra quantization unit 370.
  • FIG. 4 is a diagram for explaining a process of generating an intra basis block according to an exemplary embodiment of the present invention.
  • the block 410 is divided into a plurality of subblocks.
  • an intra basis block has a size of 4*4 pixels.
  • a block size may be determined depending on combinations of temporal and spatial scalabilities.
  • the block size may be determined using a scaling factor defined as the ratio of view layer to encoded layer. For example, when the scaling factor is 1, a block size is 16*16 pixels. When the scaling factor is 2, the block size is 32*32 pixels.
  • FIG. 5 is a diagram for explaining a process of generating an intra predicted block using the intra basis block 420 according to an exemplary embodiment of the present invention. Referring to FIG.
  • each pixel in the intra predicted block is generated using the values of pixels in the intra basis block.
  • the value of a pixel t 510 may be calculated using the values of pixel a 520, pixel b 530, pixel e 540, and pixel f 550 in the intra basis block 420.
  • the value of pixel 1 510 can be obtained by interpolating the values of neighboring pixels in an intra basis block.
  • the value of pixel t 510 is defined by Equation (2) as follows:
  • t is the value of pixel 1 510
  • a, b, e, and f are the values of pixel a 520, pixel b 530, pixel e 540, and pixel f 550, respectively
  • x and y are horizontal distances between the pixel 1 510 and the pixel a 520 and between the pixel 1 510 and the pixel b 530, respectively
  • u and v are vertical distances between the pixel t 510 and the pixel e 540 and between the pixel t and the pixel f 550, respectively.
  • a difference metric between the block (410 of FIG. 4) and the intra predicted block is provided to the determination unit (316 of FIG. 3).
  • the de ⁇ termination unit 316 uses the difference metric to determine whether to encode the block 410 in an intra predictive coding mode.
  • the intra prediction unit 314 transmits the intra predicted block to the temporal filter 320.
  • the intra prediction unit 314 predicts an intra basis block using in ⁇ formation from neighboring subblock blocks surrounding the block 410 and generate a residual intra basis block by comparing the predicted intra basis block with the previous intra basis block.
  • the intra quantization unit 370 quantizes the residual intra basis block in order to reduce the amount of information and sends the quantized residual intra basis block back to the intra prediction unit 314.
  • the intra prediction unit 314 adds the quantized residual intra basis block to the predicted intra basis block to thereby generate a new intra basis block. Then, the intra prediction unit 314 generates an intra predicted block using the new intra basis block and transmits the intra predicted block to the temporal filter 320.
  • the second exemplary embodiment offers similar performance to the first exemplary embodiment but is advantageous over the first exemplary embodiment for filtering a predicted frame in the predicted frame filtering unit 324.
  • the second exemplary embodiment also suffers less artifacts at a boundary between an inter-coded block and an intra-coded block at a low bit-rate than the first exemplary embodiment.
  • a process of predicting an intra basis block and quantizing a residual intra basis block generated with the predicted intra basis block according to the second exemplary embodiment will now be described in more detail with reference to FIG. 4.
  • the intra basis block 420 generated using representative values for subblocks in the block 410 is used to determine a mode in which the block 410 will be encoded.
  • an intra basis block is generated using information from neighboring subblocks.
  • an intra basis block for the block 410 is predicted using information from a block (subblocks) located above the block 410 ('upside block') and from a block (or subblocks) located to the left of the block 410 ('left-side block').
  • the intra basis block may be predicted according to the following rules:
  • PredictedPixel is a predicted pixel value in the intra basis block 420
  • Up- SidePixel and LeftSidePixel are respectively information from upside block and left ⁇ side block
  • DisX and DisY are respectively a distance from a pixel having a pixel valueLeftSidePixel of the left-side block and a distance from a pixel having a pixel value UpSidePixel of the upside block.
  • Up ⁇ SidePixel is 128 and LeftSidePixel is representative values of subblocks 5, 6, 7, and 8. If the representative values of subblocks 5, 6, 7, and 8 are 50, 60, 70, and 80, re ⁇ spectively, the values of pixels a, b, c, and d in the intra basis block 420 are (128*l+50*l)/(l+l), (128*2+50*l)/(2+l), (128*3+50* l)/(3+l), and ( 128*4+50* l)/(4+l), respectively.
  • pixels e, f, g, and h are (128*l+60*2)/(l+2), (128*2+60*2)/(2+2), (128*3+60*2)/ (3+2), and (128*4+60*2)/(4+l), respectively.
  • the values of pixels i, j, k, and 1 are (128*l+70*3)/(l+3), (128*2+70*3)/(2+3), (128*3+70*3)/(3+3), and (128*4+70*3)/ (4+3), respectively.
  • the values of the last four pixels m, n, o, and p are (128*l+80*4)/(l+4), (128*2+80*4)/(2+4), (128*3+80*4)/(3+4), and (128*4+80*4)/ (4+4), respectively.
  • UpSidePixel is representative values of subblocks 1, 2, 3, and 4 and LeftSidePixel is representative values of subblocks 5, 6, 7, and 8.
  • the values of pixels a, b, c, and d in the intra basis block 420 are (10*l+50*l)/(l+l), (20*2+50*l)/(2+l), (30*3+50* l)/(3+l), and (40*4+50* l)/(4+l), respectively.
  • pixels e, f, g, and h are (10*l+60*2)/(l+2), (20*2+60*2)/(2+2), (30*3+60*2)/(3+2), and (40*4+60*2)/(4+l), respectively.
  • the values of pixels i, j, k, and 1 are (10*l+70*3)/(l+3), (20*2+70*3)/(2+3), (30*3+70*3)/ (3+3), and (40*4+70*3)/(4+3), respectively.
  • the values of the last four pixels m, n, o, and p are (10*l+80*4)/(l+4), (20*2+80*4)/(2+4), (30*3+80*4)/(3+4), and (40*4+80*4)/(4+4), respectively.
  • pixel values in the intra basis block 420 can be predicted when the upside block and the left-side block are encoded in an intra predictive coding mode and in an inter predictive mode, respectively, or when the upside block and the left ⁇ side block are encoded in an inter predictive mode.
  • the pixel values in the predicted intra basis block 420 are subtracted from the pixel values in the original intra basis block to determine pixel values in a residual intra basis block.
  • the determined pixel values in the residual intra basis block may be directly subjected to quantization. However, to reduce spatial correlation, the pixel values are subjected to Hadamard transform before quantization. Quantization may be performed by a suitable quantization parameter Qp in a similar to 16*16 quantization in H.264.
  • the intra prediction unit 314 adds the quantized residual intra basis block to the intra basis block predicted using information from the neighboring subblocks and generates a new intra basis block. The intra prediction unit 314 then generates an intra predicted block by in ⁇ terpolating the new intra basis block and transmits the intra predicted block to the temporal filter 320.
  • a block is divided into 16 subblocks to generate an intra basis block
  • the block can be divided into a number of subblocks less than or greater than 16.
  • a luminance (luma) block and a chrominance (chroma) block can be divided into a different number of subblocks, respectively.
  • the luma and chroma blocks may be divided into 16 and 8 subblocks, respectively.
  • FIG. 6 is a diagram for explaining a process of filtering a predicted frame according to an exemplary embodiment of the present invention.
  • Various filtering techniques may be used to filter the values of pixels between an intra predicted block and inter predicted block. For example, when a very simple ⁇ 1, 2, 1 ⁇ filter is used, the values of pixels between the intra predicted block and the inter predicted block are determined using Equation (4):
  • Filtering can also be performed between inter predicted blocks or between intra predicted blocks.
  • FIG. 7 illustrates the process of an intra predictive coding mode according to an exemplary embodiment of the present invention.
  • a coding mode is first determined for encoding block 2 720.
  • the block 2 720 is encoded according to the following process:
  • a residual intra basis block 744 by comparing the predicted intra basis block 742 and the intra basis block 740.
  • the residual intra basis block 744 may be subjected to Hadamard transform to reduce spatial correlation.
  • the residual intra basis block 744 may be subjected to Hadamard transform to reduce spatial correlation.
  • the residual intra basis block 744 may be subjected to Hadamard transform to reduce spatial correlation.
  • the residual intra basis block 746 may be subjected to Hadamard transform to reduce spatial correlation.
  • Apply inverse quantization to the quantized residual intra basis block 746 for transmission to a decoder.
  • the inversely quantized residual intra basis block 747 is almost similar to the residual intra basis block 744 before being quantized.
  • the intra predicted block 726 is also similar to the intra predicted block 722.
  • [98] 11 Generate a residual block 728 by comparing the intra predicted block 726 with the block 2 720.
  • the residual block 728 is similar to the residual block 724.
  • [99] 12. Perform temporal filtering, wavelet transform, and quantization on the residual block 724 to generate texture information that will be contained in a bitstream.
  • FIG. 8 illustrates the process of an intra predictive coding mode according to another exemplary embodiment of the present invention. [101] For convenience of explanation, it is as sumed that coding modes for block 1 810 and block 3 830 have been already determined.
  • a coding mode is first determined for encoding block 2 820.
  • the block 2 820 is encoded according to the following process: [102] 1.
  • [104] Generate a residual block 824 by comparing the intra predicted block 822 with the block 2 820.
  • [105] 4. Determine a coding mode for the block 2 820 by comparing a cost for encoding the residual block 824 with a cost for encoding a residual block (not shown) created by inter predictive coding.
  • an intra predictive coding mode is determined as the coding mode for the block 2 820, perform temporal filtering, wavelet transform, and quantization on the residual block 824 to generate texture information that will be contained in a bitstream.
  • FIG. 9 is a block diagram of a video decoder according to an exemplary embodiment of the present invention.
  • the video decoder is assumed to decode a bitstream created by the encoding process illustrated in FIG. 7. Basically, the video decoder performs the inverse operation of an encoder on received bitstream in order to reconstruct video frames. To accomplish this, the video decoder includes a bitstream interpreter 910, an inverse quantizer 920, an inverse wavelet transformer 930, and an inverse temporal filter 940.
  • the bitstream interpreter 910 interprets a bitstream to obtain texture information, an encoded motion vector, and a quantized residual intra basis block that are then provided to the inverse quantizer 920, a motion vector decoder 950, and an inverse intra quantizer 960, respectively.
  • the quantized residual intra basis block is subjected to inverse quantization and then is added to a predicted intra basis block obtained using information from neighboring blocks, thereby generating a new intra basis block.
  • the inverse quantizer 920 inversely quantizes texture information and creates transform coefficients in the wavelet domain.
  • the inverse wavelet transformer 930 performs inverse wavelet transform on the transform coefficients to obtain a single low-pass subband and a plurality of high-pass subbands on a GOP-by-GOP basis.
  • the inverse temporal filter 940 uses the high-pass and low-pass subbands to re ⁇ construct video frames.
  • the inverse temporal filter 940 includes an inverse prediction unit 946, which receives motion vectors and residual intra basis blocks from the motion vector decoder 950 and the inverse intra quantizer 960, respectively, and re ⁇ constructs a predicted frame.
  • the inverse temporal filter 940 further includes an inverse updating unit 942.
  • the inverse temporal filter 940 further includes an inverse predicted frame filtering unit 944 filtering predicted frames obtained by an inverse prediction unit 946.
  • FIG. 9 shows a scalable video decoder
  • some of the components shown in FIG. 9 may be modified or replaced to reconstruct video frames from a bitstream produced by DCT-based encoding. Therefore, it is to be understood that the above-described exemplary em ⁇ bodiments have been provided only in a descriptive sense and will not be construed as placing any limitation on the scope of the invention.
  • a novel intra predictive coding mode reduces block artifacts introduced by video coding and improves video coding efficiency.
  • a method of filtering a predicted frame that can also be effectively used in scalable video coding to reduce the effect of block artifacts is also provided.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention concerne des procédés de codage et de décodage vidéo ainsi qu'un codeur et un décodeur vidéo. Le procédé de codage vidéo consiste à déterminer un mode de codage inter-prédictif ou un mode de codage intra-prédictif en tant que mode de codage pour chaque bloc d'une trame vidéo d'entrée, à générer une trame prédite pour la trame vidéo d'entrée sur la base de blocs prédits obtenus conformément au mode de codage déterminé, puis à coder la trame vidéo d'entrée sur la base de la trame prédite. Lorsque le mode de codage intra-prédictif est déterminé en tant que mode de codage, un bloc intra-base composé de valeurs représentatives d'un bloc est généré pour un bloc et un bloc intra-base est interpolé pour générer un bloc prédit intra pour le bloc.
PCT/KR2005/002008 2004-07-07 2005-06-27 Procedes de codage et de decodage video, codeur et decodeur video Ceased WO2006004331A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US58560404P 2004-07-07 2004-07-07
US60/585,604 2004-07-07
KR1020040055283A KR100654436B1 (ko) 2004-07-07 2004-07-15 비디오 코딩 방법과 디코딩 방법, 및 비디오 인코더와디코더
KR10-2004-0055283 2004-07-15

Publications (1)

Publication Number Publication Date
WO2006004331A1 true WO2006004331A1 (fr) 2006-01-12

Family

ID=35783097

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2005/002008 Ceased WO2006004331A1 (fr) 2004-07-07 2005-06-27 Procedes de codage et de decodage video, codeur et decodeur video

Country Status (2)

Country Link
CN (1) CN1722836A (fr)
WO (1) WO2006004331A1 (fr)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1773069A2 (fr) 2005-09-22 2007-04-11 Samsung Electronics Co, Ltd Appareil, procédé et support d'enregisterment pour le codage/décodage vidéo
EP1773068A3 (fr) * 2005-08-30 2007-06-27 Samsung Electronics Co., Ltd. Appareil, procédé et support d'enregistrement pour le codage et décodage vidéo
EP1804516A1 (fr) * 2005-08-27 2007-07-04 Samsung Electronics Co., Ltd. Procédé et dispositif de codage/décodage d'un signal vidéo
WO2009002060A1 (fr) * 2007-06-28 2008-12-31 Samsung Electronics Co., Ltd. Procédé, support et appareil destinés à coder et/ou décoder des données vidéo
WO2009002061A3 (fr) * 2007-06-27 2009-02-19 Samsung Electronics Co Ltd Procédé, support et appareil destinés à coder et/ou décoder des données vidéo
WO2011057557A1 (fr) * 2009-11-10 2011-05-19 华为技术有限公司 Procédé, appareil et système d'encodage et de décodage d'image vidéo
RU2427976C2 (ru) * 2006-07-28 2011-08-27 Кабусики Кайся Тосиба Способ и устройство для кодирования и декодирования изображения
CN103248885A (zh) * 2012-02-14 2013-08-14 乐金电子(中国)研究开发中心有限公司 帧内图像预测编解码方法及视频编解码器
US8831086B2 (en) 2008-04-10 2014-09-09 Qualcomm Incorporated Prediction techniques for interpolation in video coding
US9967590B2 (en) 2008-04-10 2018-05-08 Qualcomm Incorporated Rate-distortion defined interpolation for video coding based on fixed filter or adaptive filter
US11082697B2 (en) 2009-07-01 2021-08-03 Interdigital Vc Holdings, Inc. Methods and apparatus for signaling intra prediction for large blocks for video encoders and decoders
CN114286097A (zh) * 2021-12-22 2022-04-05 南通朝辉信息技术有限公司 二次视频编码码率控制中编码块量化增量参数优化方法
US11432001B2 (en) 2011-09-15 2022-08-30 Vid Scale, Inc. Systems and methods for spatial prediction

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102006057983A1 (de) * 2006-12-08 2008-06-12 Siemens Ag Verfahren zur Vidoecodierung einer Folge digitalisierter Bilder
KR101365575B1 (ko) * 2007-02-05 2014-02-25 삼성전자주식회사 인터 예측 부호화, 복호화 방법 및 장치
JP5188875B2 (ja) * 2007-06-04 2013-04-24 株式会社エヌ・ティ・ティ・ドコモ 画像予測符号化装置、画像予測復号装置、画像予測符号化方法、画像予測復号方法、画像予測符号化プログラム、及び画像予測復号プログラム
JP2010050860A (ja) * 2008-08-25 2010-03-04 Hitachi Ltd 画像表示装置、録画再生装置、及び画像処理方法
KR20120016980A (ko) 2010-08-17 2012-02-27 한국전자통신연구원 영상 부호화 방법 및 장치, 그리고 복호화 방법 및 장치
CN103250417B (zh) 2011-02-09 2017-06-09 Lg电子株式会社 编码和解码图像的方法及使用该方法的设备
EP3860124B8 (fr) * 2011-02-09 2025-05-07 LG Electronics Inc. Procédé de stockage d'informations de mouvement et procédé destiné à induire un prédicteur de vecteur de mouvement temporel selon ce procédé
KR101383775B1 (ko) * 2011-05-20 2014-04-14 주식회사 케이티 화면 내 예측 방법 및 장치
CN108093262B (zh) * 2011-10-24 2022-01-04 英孚布瑞智有限私人贸易公司 图像解码装置
DK2942954T3 (da) * 2011-10-24 2020-08-24 Innotive Ltd Billedafkodningsindretning
WO2015006951A1 (fr) * 2013-07-18 2015-01-22 Mediatek Singapore Pte. Ltd. Procédés permettant une décision de codeur rapide
CN104038766A (zh) * 2014-05-14 2014-09-10 三星电子(中国)研发中心 用于以图像帧为基础执行并行视频编码的装置及其方法
US10148953B2 (en) * 2014-11-10 2018-12-04 Samsung Electronics Co., Ltd. System and method for intra prediction in video coding
EP3596701A4 (fr) * 2017-03-15 2020-08-12 RealNetworks, Inc. En-têtes de blocs de codage condensés dans des systèmes et procédés de codage vidéo

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Study of Finial Committee Draft of Joint Video Specification", 5 December 2002 (2002-12-05), pages 84 - 118 *
HE Y ET AL: "H.26L-based fine granularityscalable video coding.", IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, 26 May 2002 (2002-05-26), pages 548 - 551 *
KALLURI R. ET AL: "Fine granular scalability for H.26L-based video streaming.", INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS, 18 June 2002 (2002-06-18), pages 346 - 347 *
OOSTERMANN J. ET AL: "Video Coding with H 264/AVC: tools, Performance and Complexity", IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2004, pages 7 - 28 *

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1804516A1 (fr) * 2005-08-27 2007-07-04 Samsung Electronics Co., Ltd. Procédé et dispositif de codage/décodage d'un signal vidéo
US8199815B2 (en) 2005-08-27 2012-06-12 Samsung Electronics Co., Ltd. Apparatus and method for video encoding/decoding and recording medium having recorded thereon program for executing the method
EP1773068A3 (fr) * 2005-08-30 2007-06-27 Samsung Electronics Co., Ltd. Appareil, procédé et support d'enregistrement pour le codage et décodage vidéo
EP1773069A3 (fr) * 2005-09-22 2007-06-27 Samsung Electronics Co, Ltd Appareil, procédé et support d'enregisterment pour le codage/décodage vidéo
EP1773069A2 (fr) 2005-09-22 2007-04-11 Samsung Electronics Co, Ltd Appareil, procédé et support d'enregisterment pour le codage/décodage vidéo
RU2427976C2 (ru) * 2006-07-28 2011-08-27 Кабусики Кайся Тосиба Способ и устройство для кодирования и декодирования изображения
WO2009002061A3 (fr) * 2007-06-27 2009-02-19 Samsung Electronics Co Ltd Procédé, support et appareil destinés à coder et/ou décoder des données vidéo
WO2009002060A1 (fr) * 2007-06-28 2008-12-31 Samsung Electronics Co., Ltd. Procédé, support et appareil destinés à coder et/ou décoder des données vidéo
KR101366249B1 (ko) 2007-06-28 2014-02-21 삼성전자주식회사 스케일러블 영상 부호화장치 및 방법과 그 영상 복호화장치및 방법
US8848786B2 (en) 2007-06-28 2014-09-30 Samsung Electronics Co., Ltd. Method, medium, and apparatus for encoding and/or decoding video of generating a scalable bitstream supporting two bit-depths
US11683519B2 (en) 2008-04-10 2023-06-20 Qualcomm Incorporated Rate-distortion defined interpolation for video coding based on fixed filter or adaptive filter
US8831086B2 (en) 2008-04-10 2014-09-09 Qualcomm Incorporated Prediction techniques for interpolation in video coding
US9967590B2 (en) 2008-04-10 2018-05-08 Qualcomm Incorporated Rate-distortion defined interpolation for video coding based on fixed filter or adaptive filter
US10440388B2 (en) 2008-04-10 2019-10-08 Qualcomm Incorporated Rate-distortion defined interpolation for video coding based on fixed filter or adaptive filter
US12120307B2 (en) 2009-07-01 2024-10-15 Interdigital Vc Holdings, Inc. Methods and apparatus for signaling intra prediction for large blocks for video encoders and decoders
US11936876B2 (en) 2009-07-01 2024-03-19 Interdigital Vc Holdings, Inc. Methods and apparatus for signaling intra prediction for large blocks for video encoders and decoders
US11082697B2 (en) 2009-07-01 2021-08-03 Interdigital Vc Holdings, Inc. Methods and apparatus for signaling intra prediction for large blocks for video encoders and decoders
WO2011057557A1 (fr) * 2009-11-10 2011-05-19 华为技术有限公司 Procédé, appareil et système d'encodage et de décodage d'image vidéo
US11432001B2 (en) 2011-09-15 2022-08-30 Vid Scale, Inc. Systems and methods for spatial prediction
US11785249B2 (en) 2011-09-15 2023-10-10 Vid Scale, Inc. Systems and methods for spatial prediction
US12328442B2 (en) 2011-09-15 2025-06-10 Interdigital Vc Holdings, Inc. Systems and methods for spatial prediction
CN103248885B (zh) * 2012-02-14 2018-01-26 乐金电子(中国)研究开发中心有限公司 帧内图像预测编解码方法及视频编解码器
CN103248885A (zh) * 2012-02-14 2013-08-14 乐金电子(中国)研究开发中心有限公司 帧内图像预测编解码方法及视频编解码器
CN114286097B (zh) * 2021-12-22 2023-03-17 南通朝辉信息技术有限公司 二次视频编码码率控制中编码块量化增量参数优化方法
CN114286097A (zh) * 2021-12-22 2022-04-05 南通朝辉信息技术有限公司 二次视频编码码率控制中编码块量化增量参数优化方法

Also Published As

Publication number Publication date
CN1722836A (zh) 2006-01-18

Similar Documents

Publication Publication Date Title
US20060008006A1 (en) Video encoding and decoding methods and video encoder and decoder
WO2006004331A1 (fr) Procedes de codage et de decodage video, codeur et decodeur video
US8031776B2 (en) Method and apparatus for predecoding and decoding bitstream including base layer
US7839929B2 (en) Method and apparatus for predecoding hybrid bitstream
CA2547891C (fr) Procede et appareil de codage et decodage video scalables
US8817872B2 (en) Method and apparatus for encoding/decoding multi-layer video using weighted prediction
US20060013309A1 (en) Video encoding and decoding methods and video encoder and decoder
US20060013313A1 (en) Scalable video coding method and apparatus using base-layer
US20060120450A1 (en) Method and apparatus for multi-layered video encoding and decoding
US20060013310A1 (en) Temporal decomposition and inverse temporal decomposition methods for video encoding and decoding and video encoder and decoder
AU2004302413B2 (en) Scalable video coding method and apparatus using pre-decoder
KR20070039552A (ko) 그리드 움직임 추정 및 보상을 갖는 스캘러블 비디오 코딩
AU2004310917B2 (en) Method and apparatus for scalable video encoding and decoding
WO2006118384A1 (fr) Procede et appareil destine a coder/decoder une video a couches multiples en utilisant une prediction ponderee
EP1817911A1 (fr) Procede et appareil de codage et de decodage video multicouche
AU2007221795B2 (en) Method and apparatus for scalable video encoding and decoding
WO2006006793A1 (fr) Procede de codage et decodage de video et codeur et decodeur de video
EP1766986A1 (fr) Procedes de decomposition temporelle et de decomposition temporelle inverse pour le codage et le decodage video et codeur et decodeur video
Ji et al. Architectures of incorporating MPEG-4 AVC into three dimensional subband video coding
WO2006043753A1 (fr) Procede et appareil de precodage de trains de bits hybride

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

122 Ep: pct application non-entry in european phase