WO2024256333A1 - Procédé ou appareil de codage basés sur des informations de mouvement de caméra - Google Patents
Procédé ou appareil de codage basés sur des informations de mouvement de caméra Download PDFInfo
- Publication number
- WO2024256333A1 WO2024256333A1 PCT/EP2024/065950 EP2024065950W WO2024256333A1 WO 2024256333 A1 WO2024256333 A1 WO 2024256333A1 EP 2024065950 W EP2024065950 W EP 2024065950W WO 2024256333 A1 WO2024256333 A1 WO 2024256333A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- depth
- coding block
- candidate
- parameter
- motion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
Definitions
- At least one of the present embodiments generally relates to a method or an apparatus for video encoding or decoding, and more particularly, to a method or an apparatus comprising determining, from a single depth candidate, two depth parameters of a depth model including a plane titled horizontally or vertically for a block coded with motion information representative of camera motion.
- image and video coding schemes usually employ prediction, including motion vector prediction, and transform to leverage spatial and temporal redundancy in the video content.
- prediction including motion vector prediction, and transform
- intra or inter prediction is used to exploit the intra or inter frame correlation, then the differences between the original image and the predicted image, often denoted as prediction errors or prediction residuals, are transformed, quantized, and entropy coded.
- the compressed data are decoded by inverse processes corresponding to the entropy coding, quantization, transform, and prediction.
- a method comprises video encoding by obtaining a coding block in a current image; determining a particular depth candidate in a list of depth candidates for the coding block; wherein a depth candidate allows deriving a first depth parameter of a depth model for the coding block, the depth model including a plane representative of depth values of samples of the coding block; determining a depth difference, wherein the depth difference applied to the particular depth candidate allows deriving a second depth parameter of a depth model for the coding block; determining a motion compensated prediction of the coding block with respect to a reference image from the first depth parameter and the second depth parameter using the depth model, where motion information used in motion compensation is representative of camera motion between the current image and the reference image; encoding the coding block based on the motion compensated prediction; and encoding an indication of the particular depth candidate in the list of depth candidates, an indication of the
- a method comprises video decoding by obtaining a coding block in a current image; decoding an indication of a particular depth candidate in a list of depth candidates, wherein the particular depth candidate allows deriving a first depth parameter of a depth model for the coding block, the depth model including a plane representative of depth values of samples of the coding block; decoding an indication of adepth difference; deriving a first depth parameter of a depth model for the coding block from particular depth candidate; deriving a second depth parameter of a depth model for the coding block from the particular depth candidate and the depth difference; determining a motion compensated prediction of the coding block with respect to a reference image from the first depth parameter and the second depth parameter using the depth model, where motion information used in motion compensation is representative of camera motion between the current image and the reference image; and decoding the coding block based on the motion compensated prediction.
- an apparatus comprising one or more processors, wherein the one or more processors are configured to implement the method for video encoding according to any of its variants.
- the apparatus for video encoding comprises means for implementing the method for video decoding according to any of its variants.
- another apparatus comprising one or more processors, wherein the one or more processors are configured to implement the method for video decoding according to any of its variants.
- the apparatus for video decoding comprises means for implementing the method for video decoding according to any of its variants.
- a device comprising an apparatus according to any of the decoding embodiments; and at least one of (i) an antenna configured to receive a signal, the signal including the video block, (ii) a band limiter configured to limit the received signal to a band of frequencies that includes the video block, or (iii) a display configured to display an output representative of the video block.
- a non-transitory computer readable medium containing data content generated according to any of the described encoding embodiments or variants.
- a signal comprising video data generated according to any of the described encoding embodiments or variants.
- a bitstream is formatted to include data content generated according to any of the described encoding embodiments or variants.
- a computer program product comprising instructions which, when the program is executed by a computer, cause the computer to carry out any of the described encoding/decoding embodiments or variants.
- Figure 1 illustrates a block diagram of an example apparatus in which various aspects of the embodiments may be implemented.
- Figure 2 illustrates a block diagram of an embodiment of video encoder in which various aspects of the embodiments may be implemented.
- Figure 3 illustrates a block diagram of an embodiment of video decoder in which various aspects of the embodiments may be implemented.
- Figure 4 illustrates an example texture frame of a video game with a corresponding depth map.
- Figure 5 illustrates an example architecture of a cloud gaming system.
- Figure 6 illustrates a camera motion inter tool in a codec in which various aspects of the embodiments may be implemented.
- Figure 7 illustrates four exemplary representations of a plane of a depth model according to at least one embodiment.
- Figure 8 illustrates an example of a depth model 2V to which the at least one embodiment may apply.
- Figure 9 illustrates an example of a depth model 2H to which the at least one embodiment may apply.
- Figure 10 illustrates an exemplary encoding method according to at least one embodiment.
- Figure 1 1 illustrates an exemplary encoding method according to at least one embodiment.
- Figure 12 illustrates a generic encoding method according to at least one embodiment.
- Figure 13 illustrates a generic decoding method according to at least one embodiment.
- Various embodiments relate to a video coding system in which, in at least one embodiment, it is proposed to adapt video coding tools to the cloud gaming system.
- Different embodiments are proposed hereafter, introducing some tools modifications to increase coding efficiency and improve the codec consistency when processing 2D rendered game engine video.
- an encoding method, a decoding method, an encoding apparatus, a decoding apparatus based on this principle are proposed.
- the present embodiments are presented in the context of the cloud gaming system, they may apply to any system where a 2D video may be associated to with camera parameters, such as a video captured by mobile device along with sensor’s information allowing to determine the position and characteristics of the device’s camera capturing the video. Depth information may be made available either from a sensor or other processing.
- VVC Very Video Coding
- HEVC High Efficiency Video Coding
- ECM Enhanced Compression Model
- FIG. 1 illustrates a block diagram of an example of a system in which various aspects and embodiments can be implemented.
- System 100 may be embodied as a device including the various components described below and is configured to perform one or more of the aspects described in this application. Examples of such devices, include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances, and servers.
- Elements of system 100 singly or in combination, may be embodied in a single integrated circuit, multiple ICs, and/or discrete components.
- the processing and encoder/decoder elements of system 100 are distributed across multiple ICs and/or discrete components.
- system 100 is communicatively coupled to other systems, or to other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports.
- system 100 is configured to implement one or more of the aspects described in this application.
- the system 100 includes at least one processor 110 configured to execute instructions loaded therein for implementing, for example, the various aspects described in this application.
- Processor 110 may include embedded memory, input output interface, and various other circuitries as known in the art.
- the system 100 includes at least one memory 120 (e.g. a volatile memory device, and/or a non-volatile memory device).
- System 100 includes a storage device 140, which may include non-volatile memory and/or volatile memory, including, but not limited to, EEPROM, ROM, PROM, RAM, DRAM, SRAM, flash, magnetic disk drive, and/or optical disk drive.
- the storage device 140 may include an internal storage device, an attached storage device, and/or a network accessible storage device, as non-limiting examples.
- System 100 includes an encoder/decoder module 130 configured, for example, to process data to provide an encoded video or decoded video, and the encoder/decoder module 130 may include its own processor and memory.
- the encoder/decoder module 130 represents module(s) that may be included in a device to perform the encoding and/or decoding functions. As is known, a device may include one or both of the encoding and decoding modules. Additionally, encoder/decoder module 130 may be implemented as a separate element of system 100 or may be incorporated within processor 110 as a combination of hardware and software as known to those skilled in the art.
- Program code to be loaded onto processor 110 or encoder/decoder 130 to perform the various aspects described in this application may be stored in storage device 140 and subsequently loaded onto memory 120 for execution by processor 110.
- one or more of processor 1 10, memory 120, storage device 140, and encoder/decoder module 130 may store one or more of various items during the performance of the processes described in this application. Such stored items may include, but are not limited to, the input video, the decoded video or portions of the decoded video, the bitstream, matrices, variables, and intermediate or final results from the processing of equations, formulas, operations, and operational logic.
- memory inside of the processor 1 10 and/or the encoder/decoder module 130 is used to store instructions and to provide working memory for processing that is needed during encoding or decoding.
- a memory external to the processing device (for example, the processing device may be either the processor 1 10 or the encoder/decoder module 130) is used for one or more of these functions.
- the external memory may be the memory 120 and/or the storage device 140, for example, a dynamic volatile memory and/or a non-volatile flash memory.
- an external non-volatile flash memory is used to store the operating system of a television.
- a fast external dynamic volatile memory such as a RAM is used as working memory for video coding and decoding operations, such as for HEVC, or VVC.
- the input to the elements of system 100 may be provided through various input devices as indicated in block 105.
- Such input devices include, but are not limited to, (i) an RF portion that receives an RF signal transmitted, for example, over the air by a broadcaster, (ii) a Composite input terminal, (iii) a USB input terminal, and/or (iv) an HDMI input terminal.
- the input devices of block 105 have associated respective input processing elements as known in the art.
- the RF portion may be associated with elements suitable for (i) selecting a desired frequency (also referred to as selecting a signal, or band-limiting a signal to a band of frequencies), (ii) down converting the selected signal, (iii) band-limiting again to a narrower band of frequencies to select (for example) a signal frequency band which may be referred to as a channel in certain embodiments, (iv) demodulating the down converted and bandlimited signal, (v) performing error correction, and (vi) demultiplexing to select the desired stream of data packets.
- the RF portion of various embodiments includes one or more elements to perform these functions, for example, frequency selectors, signal selectors, band-limiters, channel selectors, filters, downconverters, demodulators, error correctors, and demultiplexers.
- the RF portion may include a tuner that performs various of these functions, including, for example, down converting the received signal to a lower frequency (for example, an intermediate frequency or a near-baseband frequency) or to baseband.
- the RF portion and its associated input processing element receives an RF signal transmitted over a wired (for example, cable) medium, and performs frequency selection by filtering, down converting, and filtering again to a desired frequency band.
- Adding elements may include inserting elements in between existing elements, for example, inserting amplifiers and an analog-to-digital converter.
- the RF portion includes an antenna.
- USB and/or HDMI terminals may include respective interface processors for connecting system 100 to other electronic devices across USB and/or HDMI connections.
- various aspects of input processing for example, Reed-Solomon error correction, may be implemented, for example, within a separate input processing IC or within processor 1 10 as necessary.
- aspects of USB or HDMI interface processing may be implemented within separate interface ICs or within processor 110 as necessary.
- the demodulated, error corrected, and demultiplexed stream is provided to various processing elements, including, for example, processor 1 10, and encoder/decoder 130 operating in combination with the memory and storage elements to process the data stream as necessary for presentation on an output device.
- connection arrangement 115 for example, an internal bus as known in the art, including the I2C bus, wiring, and printed circuit boards.
- the system 100 includes communication interface 150 that enables communication with other devices via communication channel 190.
- the communication interface 150 may include, but is not limited to, a transceiver configured to transmit and to receive data over communication channel 190.
- the communication interface 150 may include, but is not limited to, a modem or network card and the communication channel 190 may be implemented, for example, within a wired and/or a wireless medium.
- Data is streamed to the system 100, in various embodiments, using a Wi-Fi network such as IEEE 802. 1 1.
- the Wi-Fi signal of these embodiments is received over the communications channel 190 and the communications interface 150 which are adapted for Wi-Fi communications.
- the communications channel 190 of these embodiments is typically connected to an access point or router that provides access to outside networks including the Internet for allowing streaming applications and other over-the-top communications.
- Other embodiments provide streamed data to the system 100 using a set-top box that delivers the data over the HDMI connection of the input block 105.
- Still other embodiments provide streamed data to the system 100 using the RF connection of the input block 105.
- the system 100 may provide an output signal to various output devices, including a display 165, speakers 175, and other peripheral devices 185.
- the other peripheral devices 185 include, in various examples of embodiments, one or more of a standalone DVR, a disk player, a stereo system, a lighting system, and other devices that provide a function based on the output of the system 100.
- control signals are communicated between the system 100 and the display 165, speakers 175, or other peripheral devices 185 using signaling such as AV. Link, CEC, or other communications protocols that enable device-to-device control with or without user intervention.
- the output devices may be communicatively coupled to system 100 via dedicated connections through respective interfaces 160, 170, and 180.
- the output devices may be connected to system 100 using the communications channel 190 via the communications interface 150.
- the display 165 and speakers 175 may be integrated in a single unit with the other components of system 100 in an electronic device, for example, a television.
- the display interface 160 includes a display driver, for example, a timing controller (T Con) chip.
- T Con timing controller
- the display 165 and speaker 175 may alternatively be separate from one or more of the other components, for example, if the RF portion of input 105 is part of a separate set-top box.
- the output signal may be provided via dedicated output connections, including, for example, HDMI ports, USB ports, or COMP outputs.
- Figure 2 illustrates an example video encoder 200, such as VVC (Versatile Video Coding) encoder.
- Figure 2 may also illustrate an encoder in which improvements are made to the VVC standard or an encoder employing technologies similar to VVC.
- the terms “reconstructed” and “decoded” may be used interchangeably, the terms “encoded” or “coded” may be used interchangeably, and the terms “image,” “picture” and “frame” may be used interchangeably.
- the term “reconstructed” is used at the encoder side while “decoded” is used at the decoder side.
- the video sequence may go through pre-encoding processing (201 ), for example, applying a color transform to the input color picture (e.g., conversion from RGB 4:4:4 to YCbCr 4:2:0), or performing a remapping of the input picture components in order to get a signal distribution more resilient to compression (for instance using a histogram equalization of one of the color components).
- Metadata can be associated with the pre-processing, and attached to the bitstream.
- a picture is encoded by the encoder elements as described below.
- the picture to be encoded is partitioned (202) and processed in units of, for example, CUs.
- Each unit is encoded using, for example, either an intra or inter mode.
- intra prediction 260
- inter mode motion estimation (275) and compensation (270) are performed.
- the encoder decides (205) which one of the intra mode or inter mode to use for encoding the unit, and indicates the intra/inter decision by, for example, a prediction mode flag.
- Prediction residuals are calculated, for example, by subtracting (210) the predicted block from the original image block.
- the prediction residuals are then transformed (225) and quantized (230).
- the quantized transform coefficients, as well as motion vectors and other syntax elements, are entropy coded (245) to output a bitstream.
- the encoder can skip the transform and apply quantization directly to the non-transformed residual signal.
- the encoder can bypass both transform and quantization, i.e. , the residual is coded directly without the application of the transform or quantization processes.
- the encoder decodes an encoded block to provide a reference for further predictions.
- the quantized transform coefficients are de-quantized (240) and inverse transformed (250) to decode prediction residuals.
- In-loop filters (265) are applied to the reconstructed picture to perform, for example, deblocking/SAO (Sample Adaptive Offset) filtering to reduce encoding artifacts.
- the filtered image is stored at a reference picture buffer (280).
- Figure 3 illustrates a block diagram of an example video decoder 300.
- a bitstream is decoded by the decoder elements as described below.
- Video decoder 300 generally performs a decoding pass reciprocal to the encoding pass as described in FIG. 2.
- the encoder 200 also generally performs video decoding as part of encoding video data.
- the input of the decoder includes a video bitstream, which can be generated by video encoder 200.
- the bitstream is first entropy decoded (330) to obtain transform coefficients, motion vectors, and other coded information.
- the picture partition information indicates how the picture is partitioned.
- the decoder may therefore divide (335) the picture according to the decoded picture partitioning information.
- the transform coefficients are de-quantized (340) and inverse transformed (350) to decode the prediction residuals. Combining (355) the decoded prediction residuals and the predicted block, an image block is reconstructed.
- the predicted block can be obtained (370) from intra prediction (360) or motion- compensated prediction (i.e., inter prediction) (375).
- In-loop filters (365) are applied to the reconstructed image.
- the filtered image is stored at a reference picture buffer (380).
- the decoded picture can further go through post-decoding processing (385), for example, an inverse color transform (e.g., conversion from YCbCr 4:2:0 to RGB 4:4:4) or an inverse remapping performing the inverse of the remapping process performed in the pre-encoding processing (201 ).
- post-decoding processing can use metadata derived in the pre-encoding processing and signaled in the bitstream.
- a video coding system such as a cloud gaming server or a device with light detection and ranging (LiDAR) capabilities may receive input video frames (e.g., texture frames) together with depth information (e.g., a depth map) and/or motion information, which may be correlated.
- input video frames e.g., texture frames
- depth information e.g., a depth map
- motion information e.g., motion information
- Figure 4 illustrates an example texture frame 402 of a video game with a corresponding depth map 404 that may be extracted (e.g., directly) from a game engine that is rendering the game scene.
- a depth map may be provided by the game engine in a floating-point representation.
- a depth map may be represented by a greylevel image, which may indicate the distance between a camera and an actual object.
- a depth map may represent the basic geometry of the captured video scene.
- a depth map may correspond to a texture picture of a video content and may include a dense monochrome picture of the same resolution as the luma picture. In examples, the depth map and the luma picture may be of different resolutions.
- Figure 5 shows an example architecture of a cloud gaming system, where a game engine may be running on a cloud server.
- the gaming system may render a game scene based on the player actions.
- the rendered game scene may be represented as a 2D video including a set of texture frames.
- the rendered game engine 2D video may be encoded into a bitstream, for example, using a video encoder.
- the bitstream may be encapsulated by a transport protocol and may be sent as a transport stream to the player’s device.
- the player’s device may de-encapsulate and decode the transport stream and present the decoded 2D video representing the game scene to the player.
- additional information such as a depth information, motion information, an object ID, an occlusion mask, camera parameters, etc.
- a game engine e.g., as outputs of the game engine
- the cloud server e.g., an encoder of the cloud
- a video to encode is generated by 3D game engine as shown in the cloud gaming system of figure 5 where the video only includes texture information and synchronized camera parameters.
- additional information e.g. depth map
- additional information e.g. depth map
- Additional information described herein such as the motion information issued from state of the art motion estimation in the encoder, or camera parameters or a combination thereof may be utilized to perform motion compensation in the rendered game engine 2D video in a video processing device (e.g., the encoder side of a video codec) as for instance described in the EP application 22306847.9 filed on 12-Dec-2022 by the same applicant which is incorporated herein by reference.
- the motion compensation generates inter prediction based on new motion information that is responsive to a new motion model in order to improve coding gains (e.g., compression gains).
- This new motion model described in the EP application 22306847.9, may render the motion of a camera in the 3D game engine.
- the processing based on the camera parameters is referred to as Camera Motion tool or Camera Motion Inter tool in the present disclosure.
- the Camera Motion tool allows predicting motion in areas of a current image where motion is only affected by the virtual camera of the game engine (its characteristics and position).
- the present principles address the deriving of depth information to be used Camera Motion Inter tool in both the encoder and the decoder by the propagation of depth parameter of neighboring blocks.
- FIG. 6 illustrates a camera motion inter tool in a codec in which various aspects of the embodiments may be implemented.
- camera motion inter tool 650 approximates the depth of the CU to be compressed to a plane.
- four different depth models can be used to approximate the plane, characterized by 1 , 2 or 3 depth parameters Pi.
- These depth parameters do not represent the depth in the 3D scene, but a third coordinate representing the depth associated to the 2D image coordinates.
- this third coordinate may be obtained from the projection of a 3D point by the game engine’s virtual camera.
- this third coordinate may be obtained from the projection of a reconstructed 3D point, the 3D point being reconstructed from motion information, such as a motion vector and reference picture, associated with the 2D image sample.
- motion information such as a motion vector and reference picture
- the depth models characterized by only few depth parameters i.e. a few motion vectors such as the motion vector of an already reconstructed sample, a depth information is available per pixel (or per 4 x4 pixel blocs in VTM) in a camera motion coded block. This depth information, associated with the camera parameters, is used to compute a motion vector per sample. Then the motion compensation 640 can be performed, as it is performed in the state of art.
- Figure 7 illustrates 4 exemplary representations of a plane of a depth model as disclosed in the EP application 22306847.9.
- the hatched planes represent some planes in the 3D game scene which are only affected by the game engine’s camera.
- an exemplary Camera Motion Coding Block (CB) 710, 720, 730 corresponding to the projection of a part of the hatched planes by the camera is represented.
- CB Camera Motion Coding Block
- a depth model for the coding block includes a plane parallel to a camera’s sensor and is characterized by one depth parameter.
- the plane 710 of the coding block may be approximated by a plane parallel to the camera’s sensor, the coding block is represented by only one depth parameter (Depth ModeH ).
- the depth parameter represents the depth value of the central sample P1 in the coding block, which is also the depth value of any sample in the coding block.
- a depth model for the coding block includes a plane 720 tilted vertically or horizontally with respect to a camera’s sensor and the depth model is characterized by two depth parameters.
- the plane may either be tilted horizontally (Depth Model 2H implying a horizontal depth interpolation) or vertically (Depth Model 2V implying a vertical depth interpolation).
- Depth Model 2V two depth parameters are required to define the depth plane.
- a first depth parameter represents a depth value of a central sample P2V- T on a top border line of the coding block and a second depth parameter represents a depth value of a central sample P2V-B on a bottom border line of the coding block.
- a first depth parameter represents a depth value of a central sample P2H-L on a left border line of the coding block and a second depth parameter represents a depth value of a central sample P2H-R on a right border line of the coding block. Then, the depth of any sample in the coding block is determined using an interpolation between the depth values indicated by the two depth parameters.
- a depth model for the coding block includes a plane tilted vertically and horizontally with respect to a camera’s sensor and the depth model is characterized by three depth parameters.
- the plane is tilted in both directions (Depth Model 3) and three parameters are required.
- the three parameters respectively represent the depth value of a topleft sample P3-TL of the coding block, a depth value of a top-right sample P3-TR of the coding block, a depth value of a bottom-left sample P3-B of the coding block.
- the positions of samples used in the depth plane model are non-limiting examples, and that the present principles may contemplate any implementation of depth parameters allowing to define the 4 plane models.
- the new Camera Motion Inter tool consists in computing the motion vectors in a new way for contents such as game engine contents.
- the Camera Motion inter tool 650 is indicated by the dotted line in encoder and decoder scheme of figure 6. Firstly, for each sample of the block (or a sub-sampled set in the block, sub-sampling by 4 in both direction for instance), an estimate of the depth of a sample is computed 620 depending on its position, the Camera Motion Depth Model and its associated parameters, where the depth in the coding block is represented with a parametric plane. Secondly, a motion vector is computed 630 depending on the sample position, the estimate depth and the camera parameters.
- the block diagram of figure 6 partially represents modules of an encoder or encoding method, for instance implemented in the exemplary encoder of figure 2.
- the block diagram of figure 6 further partially represents modules of a decoder or decoding method, for instance implemented in the exemplary decoder of figure 3.
- the Camera Motion inter tool receives some depth model parameters Pi along with camera parameters and provides motion vectors MVs used to compute the motion compensation 640.
- the camera parameters represent the characteristics and the position of the game engine’s virtual camera. They are provided for the reference image and for the current image to be encoded.
- the encoder may obtain a depth parameter Pi for a depth model i that approximates the depth at the coding block position with a plane.
- the coding block depth is approximated with a plane characterized by up to 3 parameters Pi.
- a single depth parameter P1 may be obtained by taking one of the depth of the central pixel of the coding block, an average depth around the central pixel of the coding block, or the average depth of the coding block, with or without subsampling.
- the parameter of depth model for the coding block is determined from a list of depth candidates from a causal neighborhood.
- the depth parameter of a depth candidate is derived from a depth information associated to a neighboring block previously reconstructed.
- the depth parameter of a depth candidate is derived from a motion vector information associated to a neighboring block previously reconstructed.
- An example of the deriving of an estimate depth value used as depth parameter is described in the EP application 23305419.6 filed on 28-Mar-2023 by the same applicant which is incorporated herein by reference.
- the encoder reconstructs 620 depth values of the coding block based on the depth model parameters Pi. It determines an estimation of the depth value of any sample of the coding blocks.
- a motion vector responsive to camera motion compensation per sample is computed depending on its position, its approximated depth, and the camera parameters.
- a motion vector may be computed for a block of samples.
- a motion vector is computed per block of 4x4 samples. These motion vectors are then used to perform the motion compensation 640 as known by the skilled in the art. Since this vector is computed with the depth and the camera parameters, it represents the displacement of the current sample between the reference frame and the current frame due to a camera motion (translations and/or rotations), or a modification of the camera’s characteristics (focal length, ). Different depth candidates processed by the Camera motion inter tool may be put into competition into a RDO loop to determine the motion model along with an associated depth candidate that result in the lower rate distortion cost.
- the encoder may further provide adequate signaling of the selected depth model and depth candidate to enable a decoder to recover the one or more parameters Pi to be used at the input of the camera motion inter tool 650.
- the depth parameters Pi may be derived using motion information or depth information associated to a depth candidate in a list.
- the encoder may further signal 660 camera parameters for the images.
- the Camera motion inter tool computes Camera Motion MVs as done in the encoder.
- the decoder obtains, two depth candidates in a causal neighborhood used to derive the one or more parameters Pi of the depth model of a coding block to decode where the depth model includes a plane either titled horizontally or vertically.
- the decoder obtains 680 camera parameters for the current image and for the reference image.
- the camera parameters for the reference image may be stored locally in the decoder at the reconstruction of the reference image.
- the depth parameter of the particular depth candidate may be stored locally in the decoder for processing of the next coding blocks to decode.
- the input parameters Pi characterizing the depth model of the Camera Motion inter tool represent a depth information.
- this depth information could be available at the encoder side, for instance when they are provided by a game engine as a depth map associated to the texture. But in this case, the amount of information to be transmitted to the decoder is not acceptable in the scope of video compression.
- this depth information may be obtained from motion information of the 2D video. It is desirable to provide such parameters representing the depth to the Camera Motion tool while limiting extra cost due to the transmission of depth information to the decoder and limiting the processing complexity.
- a particular depth candidates is determined for the coding block, the depth candidate along with an indication of a depth plane difference (step and sign) allows deriving two depth parameters of a depth model including a plane representative of depth values of samples of the coding block either tilted horizontally or vertically; finally a motion compensated prediction of the coding block with respect to a reference image is determined from the two depth parameters for the encoding/decoding.
- the Camera Motion Depth Model 2V and 2H only requires two depth parameters to approximate the CU to a plane. Then, this model can be used to compute a motion vector per pixel or bloc of pixels, before performing the motion compensation. For instance, when the depth is either represented by a plane tilted vertically or horizontally, the two depth parameters may be provided by two adjacent blocks coded previously as described in the application referenced under “2023P00242EP - Camera Motion inter tool with depth plane propagation” filed on the same day by the same applicant. In the present application, the two depth parameters may be provided by only one adjacent block with an indication of a depth plane difference.
- Figure 8 illustrates an example of a depth model 2V to which the at least one embodiment may apply.
- the depth parameter P2V-T defining the top depth of the CU is determined by a depth candidate among a list of depth candidates CO, C2, C4, C1 , C3 provided by the CUs located above.
- a depth candidate may have a type characterizing the type of information it comprises.
- a depth candidate may comprise motion vectors and/or depth information.
- a depth candidate may comprise a motion vector information.
- Such depth candidate may be referred to as a depth candidate of type motion vector (TYPE MV).
- a depth estimate is derived from the motion information for instance as described in the EP application 23305419.6.
- a depth candidate of type motion vector may further comprise an information associated with the spatial position of the pixel with the motion information as this spatial information is needed to estimate the depth value.
- a depth candidate may comprise a depth information.
- Such depth candidate may be referred to as a depth candidate of type depth (TYPE DEPTH).
- the bottom depth parameter P2V-B is determined among 2 values respectively resulting from a positive depth step and a negative depth step applied to the top depth parameter PV2-T.
- the encoder may compute a motion compensation and a Rate- Distortion cost to determine a particular depth candidate and the depth parameter values.
- the RD optimization only puts into competition the positive or negative value of the depth step, meaning that P2V-T is derived from the first available neighbor in an ordered list.
- the RD optimization may put into competition the different possible values for P2V-T combined with the positive or negative offset of the depth step.
- the depth step may be applied either in the depth domain or in a motion vector domain.
- an appropriate transform of the depth candidate from the motion vector domain to the depth domain or conversely may apply.
- Figure 9 illustrates an example of a depth model 2H to which the at least one embodiment may apply.
- the depth parameter P2H-L defining the left depth of the CU is determined by a depth candidate among a list of depth candidates located at the left of the current CU.
- the right depth parameter P2H-R positive and negative depth steps I are applied to the left depth candidates.
- a motion compensation and a RD cost is performed for each pair of depth parameters.
- a RD-Cost optimization process is performed for each of top or left depth value, with bottom or right depth value around these candidates.
- At least one embodiment addresses the issue of how to implement the Camera Motion tool using the Depth Model 2V or Depth Model 2H in a video codec where no depth information is available.
- the CU depth plane is tilted vertically (the depth depends on the vertical pixel position) or horizontally (the depth depends on the horizontal pixel position)
- it can be characterized by two depth parameters derived from the motion vectors of an adjacent CU (either on the top or left side of the CU) along with a depth plane difference.
- the camera motion model advantageously replaces Affine motion model to provide compression gains.
- Figure 10 illustrates an exemplary encoding method according to at least one embodiment.
- the encoder uses the Camera Motion Depth Model 2V or 2H. These two depth models represent the depth of the CU as a plane tilted vertically or horizontally, characterized by two depth parameters as shown on figure 7.
- a list of depth candidates is determined either with above neighbors for Depth Model 2V or left neighbors for Depth Model 2H.
- the type of the depth candidates (from top or left neighbors) is tested.
- the candidates in the list may be of type motion vectors TYPE MV or of type information TYPE DEPTH.
- the depth information per pixel or block of pixels may be provided as described in the application referenced under “2023P00243EP - Camera Motion inter tool with constant ClI depth” filed on the same day by the same applicant.
- the candidates in the list may be reordered to foster the depth candidate’s diversity or the candidates in the list may be pruned as described in the application referenced under “2023P00243EP - Camera Motion inter tool with constant CLI depth” as only the spatial position of the candidates in the list is distinct.
- a depth estimate is computed from a motion vector and the camera parameters as previously described in the EP application 22306847.9.
- the first step consists in building 1010 a list of candidates representing the top depth parameter P2V-T.
- the spatial position of these candidates is indicated in figure 8.
- these candidates may be depth candidates of TYPE MV or depth candidates of type TYPE DEPTH depending on the motion model (legacy inter or camera motion) associated to the CU providing the candidate.
- the type of candidate is tested in 1020. If the depth candidate is of type TYPE DEPTH, the depth candidate may be used directly as a parameter P2V-T.
- the motion vector information associated to the candidate is processed to a depth parameter.
- the positive and negative steps are applied in the motion vector domain. Therefore, if (1020) the depth candidate is of type TYPE MV, the value of the step I is directly added P2V-T+I or subtracted from the motion vector P2V-T-I in a step 1040.
- the value of the motion vector step I may for instance be declared in a table. According to a variant, different values of the motion vector step I are stored in the table. According to yet another variant, the motion vector step may be applied on the horizontal or the vertical component of the motion vector.
- a motion vector step has been applied to the motion vector value of the depth candidate, the result must be converted to depth, to provide the bottom depth P2V-B parameter.
- a bottom depth parameter P2V-B is computed from the values of the motion vectors P2V-T+I or P2V-T-I. Accordingly, applying 1040 motion vector steps I in the motion vector domain results in applying depth steps in the depth domain after the depth computation 1050. If (1020) the depth candidate is of type TYPE DEPTH, it has to be transposed back 1030 in the domain of motion vector to applying the steps I.
- the Camera Motion tool 1070 is applied to compute a motion vector per pixel or block of pixels.
- the encoder may compute a motion compensation in 1090, followed by the R-D cost computation 1092.
- This R-D cost operation 1092 is performed for each top candidate, with all the positive and negative step values, per motion vector component, to determine a particular pair of P2V-T / P2V-B parameters to be used in the encoding.
- the depth per pixel or block of pixels is stored 1080 in the codec.
- This depth information will provide depth candidates to code the next CUs.
- an encoder may signal to a decoder an index of the particular top candidate of the list of candidates to provide P2V-T, an index of the motion vector step to be applied to provide P2V-B, the sign of the motion vector step, and/or the component of the motion vector on which the step may be applied.
- the list of candidates may also comprise other candidates such as history-based candidates.
- Figure 11 illustrates an exemplary decoding method according to at least one embodiment.
- the decoding method of figure 1 1 applies the steps corresponding to the encoding method except the RD-cost.
- the particular top candidate of the list of candidates providing P2V-T, the motion vector step to be applied to provide P2V-B, the sign of the motion vector step, and/or the component of the motion vector on which the step may be applied are obtained from the decoding of dedicated syntax sent by the encoder.
- a first value of motion vector step among all possible values in the table is equal to zero.
- P2V-T or left P2V-L depth parameters are determined by depth candidates respectively located at the top or the left of the current CU, then in case the motion vector step applied to these candidates is equal to zero, P2V-B equal P2V-T and P2V-R equal P2V-L. Since the two depth parameters have the same value, the Camera Motion tool described here behaves as the Camera Motion Depth Model 1 , where the depth plane is parallel to the camera’s sensor, characterized by one depth parameter.
- the first element (table index 0) in the table of motion vector step is set to zero.
- the encoder may only signal the index of the particular depth candidate of the list of depth candidates to provide P2V- T or P2H-L, ‘0’ as the motion vector index corresponding a motion vector step equal to zero. Therefore, the Camera Motion Depth Model 1 in competition with the Camera Motion Depth Model 2V and/or 2H described in this document may be advantageously replaced by the Depth Model 2V and/or 2H with the first motion vector step equal to zero.
- the signaling of the Camera Motion Depth Model 2V and 2H with the first motion vector set to zero is a bit heavier (step index ‘0’ has to be signaled) than the signaling for Camera Motion Depth Model 1 , but the number of tools in the Codec is reduced with almost the same behavior in term of motion compensation. Besides, the ClI syntax tree is simplified and lightened, that may improve the coding efficiency.
- Figure 12 illustrates a generic encoding method 1200 according to at least one embodiment.
- a coding block in a current image is obtained.
- a particular depth candidate is determined in a list of depth candidates for a depth model of the coding block.
- the depth model includes a plane representative of depth values of samples of the coding block.
- a depth model for the coding block includes a plane tilted vertically with respect to a camera’s sensor and is characterized by the first depth parameter representative of a depth value of a sample on a top border line of the coding block and by the second depth parameter representative of a depth value of a sample on a bottom border line of the coding block.
- the list of depth candidates comprises top adjacent blocks.
- a depth model for the coding block includes a plane tilted horizontally with respect to a camera’s sensor and is characterized by the first depth parameter representative of a depth value of a sample on a left border column of the coding block and by the second depth parameter representative of a depth value of a sample on on a rigth border column of the coding block.
- the list of depth candidates comprises left adjacent blocks.
- the depth candidate allows deriving the first depth parameter of a depth model for the coding block.
- the depth candidate associated with a depth difference allows deriving the depth parameter of the depth model for the coding block.
- the encoding method determines for instance responsively to a RD-cost, a depth model to be used for the encoding along with a depth candidate and a depth difference (for instance the depth difference is represented by a motion vector step and a sign of the motion vector step).
- a motion compensated prediction of the coding block with respect to a reference image is computed from the first and second parameters.
- the motion information used in motion compensation is representative of camera motion between the current image and the reference image.
- the coding block is encoded based on the motion compensated prediction while an indication of the particular depth candidate in the list of depth candidates, and an indication of the depth difference are also encoded for signaling to the decoder.
- Figure 13 illustrates a generic decoding method 1300 according to at least one embodiment.
- a coding block to decode in a current image is obtained.
- a particular depth candidate is determined in a list of depth candidates for a depth model of the coding block from a decoded indication of the particular depth candidate signaled by the decoder.
- the depth model includes a plane representative of depth values of samples of the coding block.
- a depth model for the coding block includes a plane tilted vertically with respect to a camera’s sensor and is characterized by the first depth parameter representative of a depth value of a sample on a top border line of the coding block and by the second depth parameter representative of a depth value of a sample on a bottom border line of the coding block.
- the list of depth candidates comprises top adjacent blocks.
- a depth model for the coding block includes a plane tilted horizontally with respect to a camera’s sensor and is characterized by the first depth parameter representative of a depth value of a sample on a left border column of the coding block and by the second depth parameter representative of a depth value of a sample on a rigth border column of the coding block.
- the list of depth candidates comprises left adjacent blocks.
- the depth candidate allows deriving the first depth parameter of a depth model for the coding block.
- an indication of a depth diffrence is also decoded.
- the depth candidate associated with the depth difference allows deriving the second depth parameter of the depth model for the coding block.
- a motion compensated prediction of the coding block with respect to a reference image is computed from the first and second parameters.
- the motion information used in motion compensation is representative of camera motion between the current image and the reference image.
- the coding block is reconstructed based on the motion compensated prediction.
- each of the methods comprises one or more steps or actions for achieving the described method. Unless a specific order of steps or actions is required for proper operation of the method, the order and/or use of specific steps and/or actions may be modified or combined. Additionally, terms such as “first”, “second”, etc. may be used in various embodiments to modify an element, component, step, operation, etc., for example, a “first decoding” and a “second decoding”. Use of such terms does not imply an ordering to the modified operations unless specifically required. So, in this example, the first decoding need not be performed before the second decoding, and may occur, for example, before, during, or in an overlapping time period with the second decoding.
- modules for example, the inter prediction modules (270, 275, 375), of a video encoder 200 and decoder 300 as shown in figure 2 and figure 3.
- the present aspects are not limited to VVC or HEVC, and can be applied, for example, to other standards and recommendations, and extensions of any such standards and recommendations. Unless indicated otherwise, or technically precluded, the aspects described in this application can be used individually or in combination.
- Decoding may encompass all or part of the processes performed, for example, on a received encoded sequence in order to produce a final output suitable for display.
- processes include one or more of the processes typically performed by a decoder, for example, entropy decoding, inverse quantization, inverse transformation, and differential decoding.
- a decoder for example, entropy decoding, inverse quantization, inverse transformation, and differential decoding.
- encoding may encompass all or part of the processes performed, for example, on an input video sequence in order to produce an encoded bitstream.
- syntax elements as used herein are descriptive terms. As such, they do not preclude the use of other syntax element names.
- the implementations and aspects described herein may be implemented as various pieces of information, such as for example syntax, that can be transmitted or stored, for example.
- This information can be packaged or arranged in a variety of manners, including for example manners common in video standards such as putting the information into an SPS, a PPS, a NAL unit, a header (for example, a NAL unit header, or a slice header), or an SEI message.
- Other manners are also available, including for example manners common for system level or application level standards such as putting the information into one or more of the following:
- SDP session description protocol
- RTP Real-time Transport Protocol
- DASH MPD Media Presentation Description
- Descriptors for example as used in DASH and transmitted over HTTP, a Descriptor is associated to a Representation or collection of Representations to provide additional characteristic to the content Representation;
- RTP header extensions for example as used during RTP streaming
- ISO Base Media File Format for example as used in OMAF and using boxes which are object-oriented building blocks defined by a unique type identifier and length also known as 'atoms' in some specifications;
- HLS HTTP live Streaming
- a manifest can be associated, for example, to a version or collection of versions of a content to provide characteristics of the version or collection of versions.
- the implementations and aspects described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program).
- An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
- the methods may be implemented in, for example, an apparatus, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, for example, computers, cell phones, portable/personal digital assistants (“PDAs”), and other devices that facilitate communication of information between end-users.
- PDAs portable/personal digital assistants
- references to “one embodiment” or “an embodiment” or “one implementation” or “an implementation”, as well as other variations thereof, means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment.
- the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout this application are not necessarily all referring to the same embodiment.
- Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting the information, or retrieving the information from memory.
- Accessing the information may include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, moving the information, copying the information, calculating the information, determining the information, predicting the information, or estimating the information.
- this application may refer to “receiving” various pieces of information. Receiving is, as with “accessing”, intended to be a broad term. Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory). Further, “receiving” is typically involved, in one way or another, during operations, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
- any of the following “/”, “and/or”, and “at least one of”, for example, in the cases of “A/B”, “A and/or B” and “at least one of A and B”, is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of both options (A and B).
- such phrasing is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of the third listed option (C) only, or the selection of the first and the second listed options (A and B) only, or the selection of the first and third listed options (A and C) only, or the selection of the second and third listed options (B and C) only, or the selection of all three options (A and B and C).
- This may be extended, as is clear to one of ordinary skill in this and related arts, for as many items as are listed.
- the word “signal” refers to, among other things, indicating something to a corresponding decoder.
- the encoder signals a quantization matrix for de-quantization.
- the same parameter is used at both the encoder side and the decoder side.
- an encoder can transmit (explicit signaling) a particular parameter to the decoder so that the decoder can use the same particular parameter.
- signaling can be used without transmitting (implicit signaling) to simply allow the decoder to know and select the particular parameter. By avoiding transmission of any actual functions, a bit savings is realized in various embodiments.
- signaling can be accomplished in a variety of ways. For example, one or more syntax elements, flags, and so forth are used to signal information to a corresponding decoder in various embodiments. While the preceding relates to the verb form of the word “signal”, the word “signal” can also be used herein as a noun.
- implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
- the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
- a signal may be formatted to carry the bitstream of a described embodiment.
- Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
- the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
- the information that the signal carries may be, for example, analog or digital information.
- the signal may be transmitted over a variety of different wired or wireless links, as is known.
- the signal may be stored on a processor-readable medium.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
La présente invention concerne au moins un procédé et un appareil permettant un codage ou décodage vidéo efficace. Par exemple, des informations de mouvement sont déterminées, lesquelles sont représentatives d'un mouvement de caméra entre une image actuelle et une image de référence, l'image actuelle et l'image de référence faisant partie d'une vidéo à restitution 2D d'un moteur de jeu. Par exemple, une liste de profondeurs candidates est déterminée pour le bloc de codage, la profondeur candidate permet de déduire un premier paramètre de profondeur d'un modèle de profondeur, le modèle de profondeur incluant un plan représentatif de valeurs de profondeur d'échantillons du bloc de codage. En outre, une valeur et un signe de différence de plan sont déterminés, la profondeur candidate particulière avec la différence de plan permettant de déduire un second paramètre de profondeur. Un procédé de codage ou de décodage détermine une prédiction à compensation de mouvement du bloc de codage par rapport à une image de référence à partir des premier et second paramètres.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP23305962 | 2023-06-16 | ||
| EP23305962.5 | 2023-06-16 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2024256333A1 true WO2024256333A1 (fr) | 2024-12-19 |
Family
ID=87158511
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/EP2024/065950 Pending WO2024256333A1 (fr) | 2023-06-16 | 2024-06-10 | Procédé ou appareil de codage basés sur des informations de mouvement de caméra |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2024256333A1 (fr) |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP3657795A1 (fr) * | 2011-11-11 | 2020-05-27 | GE Video Compression, LLC | Codage multi-vues efficace utilisant une estimée de carte de profondeur et une mise à jour |
| WO2024126278A1 (fr) * | 2022-12-12 | 2024-06-20 | Interdigital Ce Patent Holdings, Sas | Procédé ou appareil de codage basé sur des informations de mouvement de caméra |
-
2024
- 2024-06-10 WO PCT/EP2024/065950 patent/WO2024256333A1/fr active Pending
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP3657795A1 (fr) * | 2011-11-11 | 2020-05-27 | GE Video Compression, LLC | Codage multi-vues efficace utilisant une estimée de carte de profondeur et une mise à jour |
| WO2024126278A1 (fr) * | 2022-12-12 | 2024-06-20 | Interdigital Ce Patent Holdings, Sas | Procédé ou appareil de codage basé sur des informations de mouvement de caméra |
Non-Patent Citations (1)
| Title |
|---|
| CHEN (QUALCOMM) Y ET AL: "Test Model 9 of 3D-HEVC and MV-HEVC", no. JCT3V-I1003, 26 August 2014 (2014-08-26), XP030132532, Retrieved from the Internet <URL:http://phenix.int-evry.fr/jct2/doc_end_user/documents/9_Sapporo/wg11/JCT3V-I1003-v1.zip JCT3V-I1003_v0.docx> [retrieved on 20140826] * |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| EP4289141A1 (fr) | Compensation spatiale d'éclairage local | |
| US20250139835A1 (en) | A method and an apparatus for encoding/decoding a 3d mesh | |
| EP4635187A1 (fr) | Procédé ou appareil de codage basé sur des informations de mouvement de caméra | |
| US20250365419A1 (en) | Methods and apparatuses for encoding/decoding a video | |
| WO2023194334A1 (fr) | Codage et décodage vidéo au moyen du rééchantillonnage d'image de référence | |
| JP2025516240A (ja) | フィルムグレインモデリングのための方法及び装置 | |
| US20230262268A1 (en) | Chroma format dependent quantization matrices for video encoding and decoding | |
| WO2024256333A1 (fr) | Procédé ou appareil de codage basés sur des informations de mouvement de caméra | |
| WO2024256339A1 (fr) | Procédé ou appareil de codage basés sur des informations de mouvement de caméra | |
| WO2024256336A1 (fr) | Procédé ou appareil de codage basés sur des informations de mouvement de caméra | |
| EP4625985A1 (fr) | Lfnst/nspt hybride explicite/implicite | |
| WO2024200466A1 (fr) | Procédé ou appareil de codage basés sur des informations de mouvement de caméra | |
| EP4625975A1 (fr) | Codage vidéo : restrictions des parametres de codage | |
| US20240397064A1 (en) | Methods and apparatuses for encoding/decoding a video | |
| US20250106428A1 (en) | Methods and apparatuses for encoding/decoding a video | |
| WO2024126279A1 (fr) | Procédé ou appareil de codage basés sur une indication d'informations de mouvement de caméra | |
| EP4668739A1 (fr) | Procédés de codage et de décodage utilisant des modes de partition géométrique et appareils correspondants | |
| US20250203092A1 (en) | Methods and apparatus for dmvr with bi-prediction weighting | |
| WO2024153634A1 (fr) | Procédé ou appareil de codage signalant une indication de paramètres de caméra | |
| WO2025114149A1 (fr) | Réordonnancement basé sur un modèle de gpm candidats | |
| WO2024213520A1 (fr) | Dérivation de mode intra basée sur un modèle à partir d'échantillons de référence décodés proches | |
| WO2025146297A1 (fr) | Procédés de codage et de décodage utilisant une prédiction intra avec des sous-partitions et appareils correspondants | |
| EP4606092A1 (fr) | Procédés et appareils de remplissage d'échantillons de référence |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24731376 Country of ref document: EP Kind code of ref document: A1 |