[go: up one dir, main page]

WO2024235608A1 - Removing some redundancies in motion information coding - Google Patents

Removing some redundancies in motion information coding Download PDF

Info

Publication number
WO2024235608A1
WO2024235608A1 PCT/EP2024/061519 EP2024061519W WO2024235608A1 WO 2024235608 A1 WO2024235608 A1 WO 2024235608A1 EP 2024061519 W EP2024061519 W EP 2024061519W WO 2024235608 A1 WO2024235608 A1 WO 2024235608A1
Authority
WO
WIPO (PCT)
Prior art keywords
motion information
predictor
initial list
difference
information predictor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
PCT/EP2024/061519
Other languages
French (fr)
Inventor
Fabrice Le Leannec
Tangi POIRIER
Franck Galpin
Pascal Le Guyadec
Antoine Robert
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
InterDigital CE Patent Holdings SAS
Original Assignee
InterDigital CE Patent Holdings SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by InterDigital CE Patent Holdings SAS filed Critical InterDigital CE Patent Holdings SAS
Publication of WO2024235608A1 publication Critical patent/WO2024235608A1/en
Pending legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/521Processing of motion vectors for estimating the reliability of the determined motion vectors or motion vector field, e.g. for smoothing the motion vector field or for correcting motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • Pixels these blocks being then partitioned into one or more sub-blocks, called original sub-blocks in the following.
  • An intra or inter prediction is then applied to each sub-block to exploit intra or inter image correlations.
  • a predictor sub-block is determined for each original sub- block.
  • a sub-block representing a difference between the original sub-block and the predictor sub-block often denoted as a prediction error sub-block, a prediction residual sub-block or simply a residual sub-block, is transformed, quantized and entropy coded to generate an encoded video stream.
  • the compressed data is decoded by inverse processes corresponding to the transform, quantization and entropic coding.
  • Inter prediction consists in predicting a current block of a current picture from at least one predictor block from a reference picture preceding or following the current picture.
  • a predictor block is identified in the reference picture by motion information.
  • motion information is now predicted using methods such Adaptive Motion Vector Prediction (AMVP) or merge.
  • AMVP Adaptive Motion Vector Prediction
  • these 2023PF00332 methods comprise a construction of a list of motion vector predictors (MVPs), a selection of a best MVP in the list and an encoding of the motion vector of the current block in the form of an index representing the selected MVP along, in some cases, a motion difference between the selected MVP and the actual motion vector of the current block.
  • one or more of the present embodiments provide a method comprising: decoding an index of a first motion information predictor in an initial list of motion information predictors for a current block of a picture; decoding a motion information difference for the current block; and, applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list.
  • determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison.
  • the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value.
  • the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process.
  • the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value.
  • the method comprises reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference.
  • one or more of the present embodiments provide a method comprising: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor.
  • determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison.
  • the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of 2023PF00332 the lowest norm of the norms of the motion information predictors of the initial list is lower than a value.
  • the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process.
  • the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value.
  • the method comprises encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference.
  • a device comprising electronic circuitry configured for: decoding an index of a first motion information predictors in an initial list of motion information predictors for a current block of a picture; decoding a motion information difference for the current block; and, applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, 2023PF00332 adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list.
  • determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison.
  • the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value.
  • the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process.
  • the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value.
  • the electronic circuitry is further configured for reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference.
  • one or more of the present embodiments provide a device comprising electronic circuitry configured for: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor.
  • determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm 2023PF00332 of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison.
  • the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value.
  • the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor.
  • adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process.
  • the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value.
  • the electronic circuitry is further configured for encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference.
  • one or more of the present embodiments provide a computer program comprising program code instructions for implementing the method according to the first or the second aspect. 2023PF00332
  • one or more of the present embodiments provide a non- transitory information storage medium storing program code instructions for implementing the method according to the first or the second aspect. 5.
  • FIG. 2 illustrates schematically an example of partitioning undergone by a picture of pixels of an original video
  • Fig.3 depicts schematically a method for encoding a video stream
  • Fig.4 depicts schematically a method for decoding an encoded video stream
  • Fig. 5A illustrates schematically an example of hardware architecture of a processing module able to implement an encoding module or a decoding module in which various aspects and embodiments are implemented
  • FIG. 5B illustrates a block diagram of an example of a first system in which various aspects and embodiments are implemented
  • Fig.5C illustrates a block diagram of an example of a second system in which various aspects and embodiments are implemented
  • Figs.6A and 6B and illustrates schematically spatial and temporal positions considered for constructing a list of merge candidates
  • Figs.7A and 7B illustrates motion coding configurations in AMVP mode
  • Fig.8A illustrates an example of motion vector reconstruction process at a decoder side of an embodiment
  • FIG. 8B illustrates an example of motion vector reconstruction process at an encoder side of an embodiment
  • Figs.9, 11 and 13 illustrate various embodiments of an encoding algorithm
  • Figs.10, 12 and 14 illustrate various embodiments of a decoding algorithm. 6.
  • DETAILED DESCRIPTION The following examples of embodiments are described in the context of a video format similar to VVC (ISO/IEC 23090-3 – MPEG-I : Versatile Video Coding (VVC) / ITU-T H.266). However, these embodiments are not limited to the video 2023PF00332 coding/decoding method corresponding to VVC.
  • Fig.1 describes an example of a context in which following embodiments can be implemented.
  • a system 11 that could be a camera, a storage device, a computer, a server or any device capable of delivering a video stream, transmits a video stream to a system 13 using a communication channel 12.
  • the video stream is either encoded and transmitted by the system 11 or received and/or stored by the system 11 and then transmitted.
  • the communication channel 12 is a wired (for example Internet or Ethernet) or a wireless (for example WiFi, 3G, 4G or 5G) network link.
  • the system 13, that could be for example a set top box, receives and decodes the video stream to generate a sequence of decoded pictures.
  • the obtained sequence of decoded pictures is then transmitted to a display system 15 using a communication channel 14, that could be a wired or wireless network.
  • the display system 15 then displays said pictures.
  • the system 13 is comprised in the display system 15.
  • Figs.2, 3 and 4 introduce an example of video format.
  • Fig.2 illustrates an example of partitioning undergone by a picture of pixels 21 of an original video sequence 20. It is considered here that a pixel is composed of three components: a luminance component and two chrominance components. Other types of pixels are however possible comprising less or more components such as only a luminance component or an additional depth component or transparency component.
  • a picture is divided into a plurality of coding entities. First, as represented by reference 23 in Fig. 2, a picture is divided in a grid of blocks called coding tree units (CTU).
  • CTU coding tree units
  • a CTU consists of an ⁇ ⁇ ⁇ block of luminance samples together with two corresponding blocks of chrominance samples.
  • N is generally a power of two having a maximum value of “128” for example.
  • a picture is divided into one or more groups of CTU. For example, it can be divided into one or more tile rows and tile columns, a tile being a sequence of CTU covering a rectangular region of a picture. In 2023PF00332 some cases, a tile could be divided into one or more bricks, each of which consisting of at least one row of CTU within the tile. Above the concept of tiles and bricks, another encoding entity, called slice, exists, that can contain at least one tile of a picture or at least one brick of a tile.
  • a CTU may be partitioned into the form of a hierarchical tree of one or more sub-blocks called coding units (CU).
  • the CTU is the root (i.e. the parent node) of the hierarchical tree and can be partitioned in a plurality of CU (i.e. child nodes).
  • Each CU becomes a leaf of the hierarchical tree if it is not further partitioned in smaller CU or becomes a parent node of smaller CU (i.e.
  • the CTU 24 is first partitioned in “4” square CU using a quadtree type partitioning.
  • the upper left CU is a leaf of the hierarchical tree since it is not further partitioned, i.e. it is not a parent node of any other CU.
  • the upper right CU is further partitioned in “4” smaller square CU using again a quadtree type partitioning.
  • the bottom right CU is vertically partitioned in “2” rectangular CU using a binary tree type partitioning.
  • the bottom left CU is vertically partitioned in “3” rectangular CU using a ternary tree type partitioning.
  • the partitioning is adaptive, each CTU being partitioned so as to optimize a compression efficiency of the CTU criterion.
  • PU prediction unit
  • TU transform unit
  • the coding entity that is used for prediction (i.e. a PU) and transform (i.e. a TU) can be a subdivision of a CU.
  • a CU of size 2 ⁇ ⁇ 2 ⁇ can be divided in PU 2411 of size ⁇ ⁇ 2 ⁇ or of size 2 ⁇ ⁇ ⁇ .
  • said CU can be divided in “4” TU 2412 of size ⁇ ⁇ ⁇ or in “16” TU of size ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • a CU comprises generally one TU and one PU.
  • the term “block” or “picture block” can be used to refer to any one of a CTU, a CU, a PU and a TU.
  • block or 2023PF00332 “picture block” can be used to refer to a macroblock, a partition and a sub-block as specified in H.264/AVC or in other video coding standards, and more generally to refer to an array of samples of numerous sizes.
  • the terms “reconstructed” and “decoded” may be used interchangeably, the terms “pixel” and “sample” may be used interchangeably, the terms “image,” “picture”, “sub-picture”, “slice” and “frame” may be used interchangeably.
  • the term “reconstructed” is used at the encoder side while “decoded” is used at the decoder side.
  • Fig.3 depicts schematically a method for encoding a video stream executed by an encoding module. Variations of this method for encoding are contemplated, but the method for encoding of Fig. 3 is described below for purposes of clarity without describing all expected variations.
  • a current original picture of an original video sequence may go through a pre-processing. For example, in a step 301, a color transform is applied to the current original picture (e.g., conversion from RGB 4:4:4 to YCbCr 4:2:0), or a remapping is applied to the current original picture components in order to get a signal distribution more resilient to compression (for instance using a histogram equalization of one of the color components).
  • the encoding of a pre-processed picture begins with a partitioning of the pre- processed picture during a step 302, as described in relation to Fig.2.
  • the pre-processed picture is thus partitioned into CTU, CU, PU, TU, etc.
  • the encoding module determines a coding mode between an intra prediction and an inter prediction.
  • the intra prediction consists of predicting, in accordance with an intra prediction method, during a step 303, the pixels of a current block from a prediction block derived from pixels of reconstructed blocks situated in a causal vicinity of the current block to be coded.
  • the result of the intra prediction is a prediction mode indicating which pixels of the blocks in the vicinity to use, and a residual block resulting from a calculation of a difference between the current block and the prediction block.
  • the inter prediction consists of predicting the pixels of a current block from a block of pixels, referred to as the reference block, of a picture preceding or following the current picture, this picture being referred to as the reference picture.
  • a block of the 2023PF00332 reference picture closest, in accordance with a similarity criterion, to the current block is determined by a motion estimation step 304.
  • a motion vector indicating the position of the reference block in the reference picture is determined.
  • Said motion vector is used during a motion compensation step 305, involving interpolation operations between samples of the reference block, to generate a prediction block.
  • a residual block is then calculated in the form of a difference between the current block and the prediction block.
  • the uni-prediction inter mode described above was the only inter mode available.
  • the family of inter modes has grown significantly and comprises now many different inter modes, such as bi-prediction modes in which a current is predicted from two reference blocks designated by two different motion information.
  • the prediction mode optimising the compression performances in accordance with a rate/distortion optimization criterion (i.e.
  • RDO criterion among the prediction modes tested (Intra prediction modes, Inter prediction modes), is selected by the encoding module.
  • the prediction mode is selected, the residual block is transformed during a step 307.
  • a plurality of type of transforms can be applied to a transformed residual block.
  • MTS Multiple Transform Selection
  • the transformed block is then quantized during a step 309. Note that the encoding module can skip the transform and apply quantization directly to the non-transformed residual signal.
  • the quantized residual block determined for the current block during an inter or intra prediction is encoded by an entropic encoder during a step 310.
  • the encoding module can bypass both transform and quantization, i.e., the entropic encoding is applied on the residual without the application of the transform or quantization processes.
  • the result of the entry coding is inserted in the video data 311.
  • the intra prediction mode is encoded by the entropic encoder during the step 310 in the video data 311.
  • 2023PF00332 When the current block is encoded according to an inter prediction, a process is applied to encode the motion information. The output of this process is then encoded by the entropic encoder during the step 310 in the video data 311.
  • AMVP Adaptive Motion Vector Prediction
  • Merge Two processes are employed to encode the motion information: AMVP (Adaptive Motion Vector Prediction) or Merge.
  • the motion information is predicted.
  • a motion vector predictor MVP
  • the MVP is selected in a list of AMVP candidates made of “2” candidates.
  • the index of the chosen MVP and the MVd are then encoded by the entropic encoder during step 310 along with the transformed and quantized residual block resulting from the inter prediction of the current block.
  • the AMVP candidate list is constructed first by deriving a first spatial candidate from a left block neighbouring the current block, if this block is available and inter coded.
  • a second spatial candidate is derived from a top block neighbouring the current block, if this block is available and inter coded.
  • a temporal candidate is derived from a so-called collocated picture of current picture, which is an already coded picture close to current picture, at a position collocated with the current block, if an inter block exist at this collocated position.
  • Each derived MVP candidate is scaled according to a temporal distance between the reference picture associated to this MVP candidate and the reference picture considered for the current block.
  • a redundancy check is then conducted between derived spatial candidates and, if a duplicate candidate exists, this candidate is discarded.
  • the final AMVP candidate list contains the two first derived MVP candidates.
  • the merge mode consists in deriving motion information of a current block from a selected motion information predictor candidate.
  • the motion information considered here includes all the inter prediction parameters of a block, that is to say: the uni-prediction or bi-prediction type, the reference picture index within each reference picture list and the motion vector(s).
  • the selected motion information predictor candidate i.e., the merge candidate
  • the selected motion information predictor candidate is selected in a list of motion information predictor candidates (i.e., in a list of merge candidates).
  • the index of the selected merge candidate is encoded.
  • the current block is considered as encoded according to a particular merge mode called skip mode.
  • the list of merge candidates is systematically made of “5” merge candidates. Up to “5” spatial positions are considered to retrieve some potential candidates for the list of merge candidates.
  • Fig. 6A illustrates schematically the five spatial positions considered for constructing a list of merge candidates. These positions are investigated according to the following order: 1. Left (A1) 2. Above (B1) 3. Above right (B0) 4. Left bottom (A0) 5. Above left (B2) Each spatial candidate is introduced in the list of merge candidates provided that the motion information corresponding to this candidate is not already present in the list of merge candidates. Then a temporal predictor noted TMVP is determined.
  • TMVP temporal predictor
  • FIG. 6B illustrates schematically collocated positions considered for determining the TMVP.
  • the determination of the TMVP consists first in investigating position H and, if no motion information is available at position H, the position C is investigated. A last pruning process is then applied to ensure that the set of spatial and temporal candidates does not contain redundant candidates.
  • candidates of another type called combined candidates, are introduced in the list of merge candidates if this list is not full.
  • zero motion vectors are introduced at the end of the merge list until it is full.
  • the whole-block-based motion representation consists in assigning one set of motion information, made of one or two motion vectors and associated reference picture(s) to an inter block.
  • the motion information of that block is represented under the form of one or two motion vectors for the whole block and a reference picture associated to each motion vector.
  • Sub-block-based motion coding mode typically consists in dividing a block into 4x4 or 8x8 luma samples subblocks and assigning an individual set of motion information (one or two couples of a motion vector and a reference picture) to each subblock.
  • HMVP History-Based Motion Vector Prediction
  • a principle of HMVP candidates is to use previously coded motion vectors as MVPs. These motion vectors are associated with adjacent or non-adjacent blocks relative to a current block. To do so, a table of HMVP candidates (i.e., HMVP table) is maintained and updated on the fly, as a first-in-first-out (FIFO) buffer of MVPs. There are up to five candidates in the HMVP table.
  • the HMVP table is updated by appending the motion information of an inter predicted block to the end of the HMVP table as a new HMVP candidate.
  • a mechanism to remove redundant HMVP candidates is applied.
  • the HMVP table is reset at each CTU row to enable parallel processing.
  • the list of merge candidates was modified and three new merge modes were introduced. The list of merge candidates is constructed with the following types of candidates: ⁇ Spatial candidates. 2023PF00332 ⁇ Temporal candidates. ⁇ HMVP candidates.
  • Pairwise Average candidates Up to one pairwise average candidate is added to the list of merge candidates. Pairwise candidates are computed as follows: the two first merge candidates present in the list of merge candidates are considered and their motion vectors are averaged. This averaging is computed separately for each reference picture list. If each of the two first merge candidates are bi-prediction ones, motion vectors related to both lists L0 and L1 are averaged. If only one motion vector is present, it is taken as is to form the pairwise candidate. 5. Zero motion vector candidate.
  • the three new merge modes comprise MMVD (Merge Mode with motion vector Difference), GPM (Geometric Partitioning Mode) and CIIP (Combined Intra/Inter Prediction). These new modes are detailed in document JVET-T2002-v2: Algorithm description for Versatile Video Coding and Test Model 11 (VTM 11), Joint Video Experts Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29, 20th Meeting, by teleconference, 7 – 16 October 2020, Jianle Chen, Yan Ye, Seung Hwan Kim. MMVD can be viewed as a kind of merge mode in which a merge candidate is refined by a MVd.
  • VTM 11 Versatile Video Coding and Test Model 11
  • JVET Joint Video Experts Team
  • a merge candidate is selected, it is further refined by a signalled MVd information.
  • the signaling of a MMVD mode comprises a merge candidate flag, an index to specify a motion magnitude and an index indicating a motion direction.
  • the merge candidate flag is signalled to specify which one is used between the first and second merge candidates.
  • the index specifying a motion magnitude and the index indicating a motion direction allow signaling a limited number of motion vector differences (MVd) on top of a signaled merge candidate, i.e., “4” vector directions and “8” magnitude values.
  • Fig. 7A illustrates a motion coding configuration in AMVP mode.
  • a block to predict is shown together with his two AMVP motion vector prediction candidates ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , and a motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ to signal. 2023PF00332 It is established that, in existing implementations of AMVP, the coding cost of a motion difference is highly correlated to the magnitude of the motion vector difference and increases as a function of the ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ magnitude.
  • Fig.7B shows that the coding configuration of the Fig.7A can also be reached by using the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ instead of ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ instead of ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , leading to a smaller motion vector difference magnitude.
  • the situation of Fig. 7A can be detected as non-optimal by a decoding module.
  • the decoder is able to detect that the use of motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is not the most optimal one to employ for pointing to the spatial position corresponding to ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ + ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , as shown in Fig. 7B.
  • the decoder is able to detect that the use of motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is the most likely one, compared to ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . Therefore, it can be considered that in the configuration of Fig. 7A, the signaling of the motion vector predictor using an index of the chosen motion vector predictor and the motion vector difference carry some redundant (or residual) information that can be exploited. Metadata such as SEI (supplemental enhancement information) messages can be attached to the encoded video stream 311.
  • SEI Supplemental Enhancement Information
  • a SEI message as defined for example in standards such as AVC, HEVC or VVC is a data container associated to a video stream and comprising metadata providing information relative to the video stream.
  • the current block is reconstructed so that the pixels corresponding to that block can be used for future predictions.
  • This reconstruction phase is also referred to as a prediction loop.
  • An inverse quantization is therefore applied to the transformed and quantized residual block during a step 312 and an inverse transformation is applied during a step 313.
  • the prediction block of the block is reconstructed.
  • the encoding module applies, when appropriate, during a step 316, a motion compensation using the motion information of the current block in order to identify each reference block of the current block.
  • the intra prediction mode is used for reconstructing the prediction block of the current block.
  • the prediction block and the 2023PF00332 reconstructed residual block are added in order to obtain the reconstructed current block.
  • an in-loop filtering intended to reduce the encoding artefacts is applied, during a step 317, to the reconstructed block.
  • In-loop filtering tools comprises deblocking filtering, SAO (Sample adaptive Offset) and ALF (Adaptive Loop Filtering).
  • SAO Sample adaptive Offset
  • ALF Adaptive Loop Filtering
  • FIG. 4 depicts schematically a method for decoding the encoded video stream 311 encoded according to method described in relation to Fig.3 executed by a decoding module. Variations of this method for decoding are contemplated, but the method for decoding of Fig. 4 is described below for purposes of clarity without describing all expected variations.
  • the decoding is done block by block. For a current block, it starts with an entropic decoding of the current block during a step 410. Entropic decoding allows to obtain, at least, the prediction mode of the block. If the block has been encoded according to an inter prediction mode, the entropic decoding allows to obtain, when appropriate, information representative of a motion of the current block and a residual block.
  • the motion information is reconstructed for the current block using the decoded information representative of the motion information.
  • the information representative of the motion comprises an index of the AMVP MVP in the AMVP list and the MVd.
  • the MVd is then added to the AMVP MVP to reconstruct the motion vector of the block.
  • an index of the merge MVP in the merge list is obtained.
  • the merge MVP corresponding to the index provides the motion information of the current block.
  • the information representative of the motion of the current block comprises some MMVD indices representative of a MVd.
  • the MVd corresponding to the MMVD 2023PF00332 indices is then added to the merge MVP to determine the motion information of the current block.
  • Steps 412, 413, 414, 415, 416 and 417 implemented by the decoding module are in all respects identical respectively to steps 412, 413, 414, 415, 416 and 417 implemented by the encoding module.
  • Decoded blocks are saved in decoded pictures and the decoded pictures are stored in a DPB 419 in a step 418.
  • the decoding module decodes a given picture, the pictures stored in the DPB 419 are identical to the pictures stored in the DPB 319 by the encoding module during the encoding of said given image.
  • the decoded picture can also be outputted by the decoding module for instance to be displayed.
  • the post-processing step 421 can comprise an inverse color transform (e.g. conversion from YCbCr 4:2:0 to RGB 4:4:4), an inverse mapping performing the inverse of the remapping process performed in the pre-processing of step 301 and a post-filtering for improving the reconstructed pictures based for example on filter parameters provided in a SEI message.
  • Fig. 5A illustrates schematically an example of hardware architecture of a processing module 500 able to implement an encoding module or a decoding module capable of implementing respectively a method for encoding of Fig.3 and a method for decoding of Fig. 4 modified according to different aspects and embodiments.
  • the encoding module is for example comprised in the system 11 when this apparatus is in charge of encoding the video stream.
  • the decoding module is for example comprised in the system 13.
  • the processing module 500 comprises, connected by a communication bus 5005: a processor or CPU (central processing unit) 5000 encompassing one or more microprocessors, general purpose computers, special purpose computers, and processors based on a multi-core architecture, as non-limiting examples; a random access memory (RAM) 5001; a read only memory (ROM) 5002; a storage unit 5003, which can include non-volatile memory and/or volatile memory, including, but not limited to, Electrically Erasable Programmable Read-Only Memory (EEPROM), Read-Only Memory (ROM), Programmable Read-Only Memory (PROM), Random Access Memory (RAM), Dynamic Random Access Memory (DRAM), Static Random Access Memory (SRAM), flash, magnetic disk drive, and/or 2023PF00332 optical disk drive, or a storage medium reader, such as a SD (secure digital)
  • the communication interface 5004 can include, but is not limited to, a transceiver configured to transmit and to receive data over a communication channel.
  • the communication interface 5004 can include, but is not limited to, a modem or network card. If the processing module 500 implements a decoding module, the communication interface 5004 enables for instance the processing module 500 to receive encoded video streams and to provide a sequence of decoded pictures. If the processing module 500 implements an encoding module, the communication interface 5004 enables for instance the processing module 500 to receive a sequence of original picture data to encode and to provide an encoded video stream.
  • the processor 5000 is capable of executing instructions loaded into the RAM 5001 from the ROM 5002, from an external memory (not shown), from a storage medium, or from a communication network.
  • the processor 5000 When the processing module 500 is powered up, the processor 5000 is capable of reading instructions from the RAM 5001 and executing them. These instructions form a computer program causing, for example, the implementation by the processor 5000 of a decoding method as described in relation with Fig.4, an encoding method described in relation to Fig.3, and methods described in relation to Figs.6 or 7, these methods comprising various aspects and embodiments described below in this document. All or some of the algorithms and steps of the methods of Figs.
  • 4, 6 and 7 may be implemented in software form by the execution of a set of instructions by a programmable machine such as a DSP (digital signal processor) or a microcontroller, or be implemented in hardware form by a machine or a dedicated component such as a FPGA (field-programmable gate array) or an ASIC (application-specific integrated circuit).
  • a programmable machine such as a DSP (digital signal processor) or a microcontroller
  • FPGA field-programmable gate array
  • ASIC application-specific integrated circuit
  • microprocessors, general purpose computers, special purpose computers, processors based or not on a multi-core architecture, DSP, microcontroller, FPGA and ASIC are electronic circuitry adapted to implement at least partially the methods of Figs.3, 4, 6 and 7.
  • Fig. 5C illustrates a block diagram of an example of the system 13 in which 2023PF00332 various aspects and embodiments are implemented.
  • the system 13 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances and head mounted display. Elements of system 13, singly or in combination, can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components.
  • the system 13 comprises one processing module 500 that implements a decoding module.
  • system 13 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports. In various embodiments, the system 13 is configured to implement one or more of the aspects described in this document.
  • the input to the processing module 500 can be provided through various input modules as indicated in block 531.
  • Such input modules include, but are not limited to, (i) a radio frequency (RF) module that receives an RF signal transmitted, for example, over the air by a broadcaster, (ii) a component (COMP) input module (or a set of COMP input modules), (iii) a Universal Serial Bus (USB) input module, and/or (iv) a High Definition Multimedia Interface (HDMI) input module.
  • RF radio frequency
  • COMP component
  • USB Universal Serial Bus
  • HDMI High Definition Multimedia Interface
  • Other examples not shown in FIG.5C, include composite video.
  • the input modules of block 531 have associated respective input processing elements as known in the art.
  • the RF module can be associated with elements suitable for (i) selecting a desired frequency (also referred to as selecting a signal, or band-limiting a signal to a band of frequencies), (ii) down-converting the selected signal, (iii) band-limiting again to a narrower band of frequencies to select (for example) a signal frequency band which can be referred to as a channel in certain embodiments, (iv) demodulating the down-converted and band- limited signal, (v) performing error correction, and (vi) demultiplexing to select the desired stream of data packets.
  • a desired frequency also referred to as selecting a signal, or band-limiting a signal to a band of frequencies
  • down-converting the selected signal for example
  • band-limiting again to a narrower band of frequencies to select (for example) a signal frequency band which can be referred to as a channel in certain embodiments
  • demodulating the down-converted and band- limited signal (v) performing error correction, and (vi) demultiplexing to select the desired stream
  • the RF module of various embodiments includes one or more elements to perform these functions, for example, frequency selectors, signal selectors, band-limiters, channel selectors, filters, downconverters, demodulators, error correctors, and demultiplexers.
  • the RF portion can include a tuner that performs 2023PF00332 various of these functions, including, for example, down-converting the received signal to a lower frequency (for example, an intermediate frequency or a near-baseband frequency) or to baseband.
  • the RF module and its associated input processing element receives an RF signal transmitted over a wired (for example, cable) medium, and performs frequency selection by filtering, down- converting, and filtering again to a desired frequency band.
  • Adding elements can include inserting elements in between existing elements, such as, for example, inserting amplifiers and an analog-to-digital converter.
  • the RF module includes an antenna.
  • the USB and/or HDMI modules can include respective interface processors for connecting system 13 to other electronic devices across USB and/or HDMI connections. It is to be understood that various aspects of input processing, for example, Reed-Solomon error correction, can be implemented, for example, within a separate input processing IC or within the processing module 500 as necessary. Similarly, aspects of USB or HDMI interface processing can be implemented within separate interface ICs or within the processing module 500 as necessary.
  • the demodulated, error corrected, and demultiplexed stream is provided to the processing module 500.
  • Various elements of system 13 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards.
  • I2C Inter-IC
  • the processing module 500 is interconnected to other elements of said system 13 by the bus 5005.
  • the communication interface 5004 of the processing module 500 allows the system 13 to communicate on the communication channel 12.
  • the communication channel 12 can be implemented, for example, within a wired and/or a wireless medium.
  • Wi-Fi Wireless Fidelity
  • IEEE 802.11 IEEE refers to the Institute of Electrical and Electronics Engineers
  • the Wi- 2023PF00332 Fi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications.
  • the communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications.
  • Other embodiments provide streamed data to the system 13 using the RF connection of the input block 531. As indicated above, various embodiments provide data in a non- streaming manner.
  • various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network.
  • the system 13 can provide an output signal to various output devices, including the display system 15, speakers 56, and other peripheral devices 57.
  • the display system 15 of various embodiments includes one or more of, for example, a touchscreen display, an organic light-emitting diode (OLED) display, a curved display, and/or a foldable display.
  • the display 15 can be for a television, a tablet, a laptop, a cell phone (mobile phone), a head mounted display or other devices.
  • the display system 15 can also be integrated with other components (for example, as in a smart phone), or separate (for example, an external monitor for a laptop).
  • the other peripheral devices 57 include, in various examples of embodiments, one or more of a stand-alone digital video disc (or digital versatile disc) (DVR, for both terms), a disk player, a stereo system, and/or a lighting system.
  • Various embodiments use one or more peripheral devices 57 that provide a function based on the output of the system 13. For example, a disk player performs the function of playing an output of the system 13.
  • control signals are communicated between the system 13 and the display system 15, speakers 56, or other peripheral devices 57 using signaling such as AV.Link, Consumer Electronics Control (CEC), or other communications protocols that enable device-to-device control with or without user intervention.
  • the output devices can be communicatively coupled to system 13 via dedicated connections through respective interfaces 532, 533, and 534. Alternatively, the output devices can be connected to system 13 using the communications channel 12 via the communications interface 5004 or a dedicated communication channel via the communication interface 5004.
  • the display system 15 and speakers 56 can be integrated in a single unit with the other components of system 13 in an electronic device such as, for example, a television.
  • the display interface 2023PF00332 532 includes a display driver, such as, for example, a timing controller (T Con) chip.
  • T Con timing controller
  • the display system 15 and speaker 56 can alternatively be separate from one or more of the other components.
  • Fig. 5C illustrates a block diagram of an example of the system 11 in which various aspects and embodiments are implemented.
  • System 11 is very similar to system 13.
  • the system 11 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, a camera and a server.
  • Elements of system 11, singly or in combination, can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components.
  • IC integrated circuit
  • the system 11 comprises one processing module 500 that implements an encoding module.
  • the system 11 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports.
  • the system 11 is configured to implement one or more of the aspects described in this document.
  • the input to the processing module 500 can be provided through various input modules as indicated in block 531 already described in relation to Fig.5C.
  • Various elements of system 11 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards.
  • I2C Inter-IC
  • the processing module 500 is interconnected to other elements of said system 11 by the bus 5005.
  • the communication interface 5004 of the processing module 500 allows the system 500 to communicate on the communication channel 12.
  • Data is streamed, or otherwise provided, to the system 11, in various embodiments, using a wireless network such as a Wi-Fi network, for example IEEE 802.11 (IEEE refers to the Institute of Electrical and Electronics Engineers).
  • IEEE 802.11 IEEE refers to the Institute of Electrical and Electronics Engineers.
  • the Wi- 2023PF00332 Fi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications.
  • the communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications.
  • Other embodiments provide streamed data to the system 11 using the RF connection of the input block 531.
  • various embodiments provide data in a non-streaming manner.
  • various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network.
  • the data provided to the system 11 can be provided in different format.
  • these data are encoded and compliant with a known video compression format such as AV1, VP9, VVC, HEVC, AVC, etc.
  • these data are raw data provided for example by a picture and/or audio acquisition module connected to the system 11 or comprised in the system 11. In that case, the processing module take in charge the encoding of these data.
  • the system 11 can provide an output signal to various output devices capable of storing and/or decoding the output signal such as the system 13.
  • decoding can encompass all or part of the processes performed, for example, on a received encoded video stream in order to produce a final output suitable for display.
  • processes include one or more of the processes typically performed by a decoder, for example, entropy decoding, inverse quantization, inverse transformation, and prediction.
  • processes also, or alternatively, include processes performed by a decoder of various implementations described in this application, for example, for reconstruction motion information.
  • decoding process is intended to refer specifically to a subset of operations or generally to the broader decoding process will be clear based on the context of the specific descriptions and is believed to be well understood by those skilled in the art.
  • Various implementations involve encoding.
  • encoding as used in this application can encompass all or part of the processes performed, for example, on an input video sequence in order to 2023PF00332 produce an encoded video stream.
  • processes include one or more of the processes typically performed by an encoder, for example, partitioning, prediction, transformation, quantization, and entropy encoding.
  • such processes also, or alternatively, include processes performed by an encoder of various implementations described in this application, for example, for predicting motion information.
  • encoding process is intended to refer specifically to a subset of operations or generally to the broader encoding process will be clear based on the context of the specific descriptions and is believed to be well understood by those skilled in the art.
  • syntax elements names as used herein are descriptive terms. As such, they do not preclude the use of other syntax element names.
  • Various embodiments refer to rate distortion optimization.
  • the rate distortion optimization is usually formulated as minimizing a rate distortion function, which is a weighted sum of the rate and of the distortion.
  • the approaches may be based on an extensive testing of all encoding options, including all considered modes or coding parameters values, with a complete evaluation of their coding cost and related distortion of a reconstructed signal after coding and decoding.
  • Faster approaches may also be used, to save encoding complexity, in particular with computation of an approximated distortion based on a prediction or a prediction residual signal, not the reconstructed one.
  • Mix of these two approaches can also be used, such as by using an approximated distortion for only some of the possible encoding options, and a complete distortion for other encoding options.
  • Other approaches only evaluate a subset of the possible encoding options. More generally, many approaches employ any of a variety of techniques to perform the optimization, but the optimization is not necessarily a complete evaluation of both the coding cost and related distortion. 2023PF00332
  • the implementations and aspects described herein can be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal.
  • An apparatus can be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods can be implemented, for example, in a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device.
  • processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs"), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • references to “one embodiment” or “an embodiment” or “one implementation” or “an implementation”, as well as other variations thereof, means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment.
  • the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout this application are not necessarily all referring to the same embodiment. Additionally, this application may refer to “determining” various pieces of information. Determining the information can include one or more of, for example, estimating the information, calculating the information, predicting the information, retrieving the information from memory or obtaining the information for example from another device, module or from user.
  • this application may refer to “accessing” various pieces of information. Accessing the information can include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, moving the information, copying the information, calculating the information, determining the information, predicting the information, or estimating the information. Additionally, this application may refer to “receiving” various pieces of information. Receiving is, as with “accessing”, intended to be a broad term. Receiving the information can include one or more of, for example, accessing the information, or 2023PF00332 retrieving the information (for example, from memory).
  • “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information. It is to be appreciated that the use of any of the following “/”, “and/or”, and “at least one of”, “one or more of” for example, in the cases of “A and/or B” and “at least one of A and B”, “one or more of A and B” is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of both options (A and B).
  • the word “signal” refers to, among other things, indicating something to a corresponding decoder.
  • the encoder signals a use of some coding tools.
  • the same parameters can be used at both the encoder side and the decoder side.
  • an encoder can transmit (explicit signaling) a particular parameter to the decoder so that the decoder can use the same particular parameter.
  • signaling can be used without transmitting (implicit signaling) to simply allow the decoder to know and select the particular parameter. By avoiding transmission of any actual functions, a bit savings is realized in various embodiments.
  • signaling can be accomplished in a variety of ways. For example, one or more syntax elements, flags, and so forth are used to signal information to a corresponding decoder in various embodiments. While the preceding relates to the verb form of the word “signal”, the word “signal” can also be used herein as a noun. 2023PF00332
  • implementations can produce a variety of signals formatted to carry information that can be, for example, stored or transmitted. The information can include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal can be formatted to carry the encoded video stream and SEI messages of a described embodiment.
  • Such a signal can be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting can include, for example, encoding an encoded video stream and modulating a carrier with the encoded video stream.
  • the information that the signal carries can be, for example, analog or digital information.
  • the signal can be transmitted over a variety of different wired or wireless links, as is known.
  • the signal can be stored on a processor-readable medium.
  • the following embodiments try to take advantage of the redundant (or residual) information existing in the encoding of the motion information in the form of an index on a MVP and a MVd.
  • Fig. 8A illustrates an example of motion vector reconstruction process at decoder side of an embodiment.
  • a motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ as in Fig. 8A is signaled in video data along with a syntax element mvp_flag indicating a use of motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ as the motion vector predictor of a current block.
  • a decoding-side validity checking process is applied to determine a validity of the combination of ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ (represented by the index mvp_flag) and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • the validity checking process is based on the following decoding-side criterion: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • This decoding side criterion is easy to compute and is strictly equivalent to the following second decoding side criterion: 2023PF00332 ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the AMVP list is adapted to obtain an adapted list.
  • the adaptation of the AMVP list consists in replacing the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ by an alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ , so that the combination of ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • Spatial position ⁇ ⁇ ⁇ shown by Fig.8A. corresponds to the spatial position pointed to by the adaptive motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ .
  • the adaptive motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ .
  • the motion vector difference required to reach the target spatial position ( ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ starting from spatial position ⁇ ⁇ ⁇ is of smaller magnitude than the MVD required when starting from spatial position ⁇ ⁇ , hence leads to lower MVD coding rate cost.
  • the region ⁇ ⁇ is the set of spatial positions that are reached from spatial position ⁇ ⁇ with a smaller MVD than from spatial position ⁇ ⁇ .
  • the proposed embodiment leads to the establishment of a further spatial region ⁇ ⁇ ⁇ , shown on Fig.8A.
  • Region ⁇ ⁇ ⁇ correspond to spatial positions that can be reached from position ⁇ ⁇ ⁇ with a smaller MVD in magnitude than from both positions ⁇ ⁇ and ⁇ ⁇ .
  • coding a motion vector pointing to a position in ⁇ ⁇ ⁇ is achieved with 2023PF00332 smaller MVD magnitudes than in prior methods, hence reduced MVD coding cost.
  • increased compression is expected from the proposed embodiment.
  • the decoding-side validity checking process is used together with a compliant encoding-side validity checking process illustrated by Fig.8B.
  • a ssume a motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is to be encoded in video data for a current block.
  • Motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ points to a spatial position ⁇ as shown in Fig. 8B.
  • Position ⁇ is closer to position ⁇ pointed to by ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ than position ⁇ pointed to by ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . .
  • the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is selected to predict the motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and a motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is computed between ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • An encoding side validity checking process checks if the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ can be modified to produce a smaller motion vector difference magnitude, and in a way detectable by a decoder.
  • the encoding-side validity checking process comprises checking the following encoding side criterion: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • the encoding module replaces the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ by an alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ 1 ⁇ ⁇ 2 ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ 1 ⁇ ⁇ 2 ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • has value ⁇ ⁇ .
  • the alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ is computed as follows: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the encoding algorithm is for example executed by the processing module 500 of the system 11 when the system 11 implements the encoding module of Fig. 3.
  • the encoding algorithm is for example executed during step 308.
  • the processing module 500 obtains a motion vector predictor for a motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ of a current block (obtained in step 304).
  • the processing module 500 selects the motion vector predictor minimizing the motion vector difference among the motion vector predictor candidates ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ of the AMVP list.
  • the selected motion vector predictor candidate noted ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the validity checking process comprises determining from the encoding side criterion of equation Eq. 3 that the non-selected motion vector predictor of the AMVP list ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and the motion information difference form a valid combination: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • step 903 the processing module 500 computes the encoding side criterion.
  • step 904 the processing module 500 determines if the encoding side criterion 2023PF00332 is verified.
  • step 904 is followed by a step 909.
  • step 909 the processing module 500 encodes the motion vector predictor index idx in the video data 311.
  • step 910 the processing module 500 encodes the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ in the video data 311.
  • step 904 is followed by a step 905.
  • the processing module 500 adapt the AMVP list to obtain an adapted list.
  • the adapting of the AMVP list comprises replacing the motion vector predictor candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ by an alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the recomputed motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ is computed as follows: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • a step 908 the processing module 500 encodes the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ in the video data 311.
  • Fig.10 illustrates an embodiment of a decoding algorithm.
  • the decoding algorithm is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of Fig. 4.
  • the decoding algorithm is for example executed during step 408.
  • the processing module decodes an index ⁇ ⁇ ⁇ ⁇ of a motion vector predictor for a current block from the video data 311.
  • This index ⁇ ⁇ ⁇ ⁇ ⁇ identifies a motion vector predictor among the motion vector predictor candidates ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ of the AMVP list. 2023PF00332
  • the processing module 500 decodes a motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ for the current block from the video data 311.
  • the processing module 500 applies a decoding-side validity checking process.
  • the decoding-side validity checking process comprises determining from the decoding-side criterion of equation Eq.
  • step 1003 the processing module 500 computes the decoding side criterion: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ 1 ⁇ ⁇ ⁇ 2 ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the processing module 500 reconstructs the motion vector of the current block ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ as the sum between the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the processing module 500 adapt the AMVP list to obtain an adapted list.
  • the adaptation of the AMVP list comprises replacing the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ by an alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • an additional decoding-side criterion is computed during step 1005 to determine if the alternative motion vector predictor is the one computed in step 1005 (i.e., ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ) or is the zero-motion- vector.
  • the additional decoding-side criterion consists in checking the following condition: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . min ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • step 1005 the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ is replaced by the zero-motion vector.
  • an additional encoding-side criterion identical to the additional decoding-side criterion is applied by the processing module 500 of the system 11 in step 905 to determine if the alternative motion vector predictor is the one computed in step 905 (i.e., ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ) or is the zero-motion-vector.
  • the decoding-side criterion of equation Eq.1 is used by the encoding module and the decoding module.
  • Equation Eq.1 the decoding-side criterion of equation Eq.1 is called validity criterion in the following and is used in its generalized form represented by equation Eq.5 below: ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . .
  • Equation Eq. 5 the encoding and the decoding modules consider a next candidate in the succession of potential candidates considered during the AMVP and/or merge candidates list construction process. For instance, w.r.t. AMVP, spatial candidates, a temporal motion vector candidate, HMVP candidates and then the zero-motion candidate may be considered during the AMVP candidate list construction process.
  • potential candidates considered in the process for constructing a list of motion vector predictors are used to 2023PF00332 generate a further motion vector predictor, noted ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • the second variant allows to diversify the set of potential motion vector predictors that can be used in combination with a given motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • Fig.11 illustrates an embodiment of an encoding algorithm compliant with the second variant.
  • the encoding algorithm of Fig. 11 is for example executed by the processing module 500 of the system 11 when the system 11 implements the encoding module of Fig.3.
  • the encoding algorithm of Fig.11 is for example executed during step 308. Steps 901, 902, 909 and 910 remains unchanged compared to the algorithm of Fig.9.
  • the processing module 500 applies a validity checking process.
  • the validity checking process comprises determining from the validity criterion of equation Eq.
  • step 1101 the processing module 500 determines if the validity criterion is verified.
  • step 1101 is followed by a step 1102.
  • the processing module 500 tries to obtain a next candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ from the potential candidates considered during the AMVP and/or merge candidates list construction process.
  • the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 909. Otherwise, step 1103 is followed by a step 1105.
  • the processing module 500 recomputes a motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ between the 2023PF00332 motion vector ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ of a current block and the next candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • Equation Eq. 6 correspond to equation Eq. 5 adapted to the combination of the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ and the non-selected motion vector predictor of the AMVP list ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • a step 1107 the processing module 500 determines if the validity criterion is verified. I f the combination of ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ and the motion information difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ is valid (i.e., if the condition of the above criterion of equation Eq.5 is not verified), step 1107 is followed by a step 909.
  • step 1107 the processing module 500 compares the magnitude of the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ to the magnitude of the motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • step 1108 is followed by step 909. If ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , step 1108 is followed by steps 907 and 908 already explained in relation to Fig.9.
  • Fig. 12 illustrates an embodiment of a decoding algorithm compliant with the second variant. The encoding algorithm of Fig.
  • step 12 is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of Fig.4.
  • the encoding algorithm of Fig.13 is for example executed during step 408.
  • Steps 1001, 1002, 1003, 1004 and 1007 remains unchanged compared to the algorithm of Fig.10. If the combination is not valid (i.e. if the validity criterion of equation Eq. 5 is 2023PF00332 fulfilled), step 1004 is followed by a step 1201.
  • the processing module 500 tries to obtain a next candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ from the potential candidates considered during the AMVP and/or merge candidates list construction process.
  • the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 1007. If a next candidate can be obtained, the processing module 500 applies a step 1204.
  • step 1204 the processing module 500 adapts the AMVP list to obtain an adapted list.
  • the adaptation of the AMVP list comprises replacing the motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ by an alternative motion vector predictor ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ : ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ′ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the validity criterion is used by the encoding module and the decoding module.
  • the encoding and the decoding modules evaluate potential candidates considered during the AMVP and/or merge candidates list construction process (for instance, w.r.t. AMVP, spatial candidates, a temporal motion vector candidate, HMVP candidates and then the zero-motion) until finding a candidate sufficiently different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ .
  • this identified candidate is used to generate a further motion vector predictor, noted ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ , different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • the third variant allows to increase the diversity of potential motion vector predictors that can be used in combination with a given motion vector difference ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ .
  • Fig.13 illustrates an embodiment of an encoding algorithm compliant with the third variant.
  • the encoding algorithm of Fig. 13 is for example executed by the processing 2023PF00332 module 500 of the system 11 when the system 11 implements the encoding module of Fig.3.
  • the encoding algorithm of Fig.13 is for example executed during step 308. Steps 901, 902, 1100, 1101, 909 and 910 remains unchanged compared to the same steps in the algorithms of Figs.9 and 11.
  • step 1101 is followed by an iterative process comprising steps 1102, 1103 and a new step 1104 allowing evaluating the potential candidates considered during the AMVP and/or merge candidates list construction process. Potential candidates are successively considered (i.e. run though) in the order in which they are considered for the construction of the AMVP and/or the merge list.
  • the processing module 500 tries to obtain a next candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . ⁇ .
  • step 1104 is followed by steps 1105, 1106, as already explained in relation to Fig.11.
  • Fig. 14 illustrates an embodiment of a decoding algorithm compliant with the third variant. The encoding algorithm of Fig.
  • step 14 is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of 2023PF00332 Fig.4.
  • the encoding algorithm of Fig.13 is for example executed during step 408. Again, steps 1001, 1002, 1003, 1004 and 1007 remains unchanged compared to the algorithm of Fig.10. If the combination is not valid (i.e., if the validity criterion of equation Eq.5 is fulfilled), step 1004 is followed by an iterative process comprising steps 1201, 1202 and a new step 1203 allowing evaluating the potential candidates considered during the AMVP and/or merge candidates list construction process. Potential candidates are run though in the order in which they are considered for the construction of the AMVP and/or the merge list.
  • step 1201 the processing module 500 tries to obtain a next candidate ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇
  • the processing module 500 determines if the obtained candidate is sufficiently different from ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ . .
  • step 1203 is followed by steps 1204 and 1006 as already Fig.12.
  • steps 1204 and 1006 we described above a number of embodiments. Features of these embodiments can be provided alone or in any combination.
  • embodiments can include one or more of the following features, devices, or aspects, alone or in any combination, across various claim categories and types: 2023PF00332 ⁇ A bitstream or signal that includes one or more of the described syntax elements, or variations thereof. ⁇ Creating and/or transmitting and/or receiving and/or decoding a bitstream or signal that includes one or more of the described syntax elements, or variations thereof. ⁇ A TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described. ⁇ A TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described, and that displays (e.g. using a monitor, screen, or other type of display) a resulting picture.
  • a TV, set-top box, cell phone, tablet, or other electronic device that tunes (e.g. using a tuner) a channel to receive a signal including an encoded video stream, and performs at least one of the embodiments described.
  • a TV, set-top box, cell phone, tablet, or other electronic device that receives (e.g. using an antenna) a signal over the air that includes an encoded video stream, and performs at least one of the embodiments described.
  • a server, camera, cell phone, tablet or other electronic device that tunes (e.g. using a tuner) a channel to transmit a signal including an encoded video stream, and performs at least one of the embodiments described.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method comprising: decoding (1001) an index of a first motion information predictor in an initial list of motion information predictors for a current block of a picture; decoding (1002) a motion information difference for the current block; and, applying (1003, 1004) a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, adapting (1005) the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process.

Description

2023PF00332 REMOVING SOME REDUNDANCIES IN MOTION INFORMATION CODING 1. CROSS REFERENCE TO RELATED APPLICATIONS This application claims priority to European Application No.23305765.2, filed May 15, 2023 and to European Application No. 24305047.3 filed January 9, 2024, which are incorporated herein by reference in their entirety. 2. TECHNICAL FIELD At least one of the present embodiments generally relates to a method and a device for improving encoding of motion information in video compression methods. 3. BACKGROUND To achieve high compression efficiency, video coding schemes usually employ predictions and transforms to leverage spatial and temporal redundancies in a video content. During an encoding, pictures of the video content are divided into blocks of samples (i.e. Pixels), these blocks being then partitioned into one or more sub-blocks, called original sub-blocks in the following. An intra or inter prediction is then applied to each sub-block to exploit intra or inter image correlations. Whatever the prediction method used (intra or inter), a predictor sub-block is determined for each original sub- block. Then, a sub-block representing a difference between the original sub-block and the predictor sub-block, often denoted as a prediction error sub-block, a prediction residual sub-block or simply a residual sub-block, is transformed, quantized and entropy coded to generate an encoded video stream. To reconstruct the video, the compressed data is decoded by inverse processes corresponding to the transform, quantization and entropic coding. Inter prediction consists in predicting a current block of a current picture from at least one predictor block from a reference picture preceding or following the current picture. A predictor block is identified in the reference picture by motion information. As the video compression methods evolve, the coding of the motion information has been greatly improved. In particular, motion information is now predicted using methods such Adaptive Motion Vector Prediction (AMVP) or merge. Basically, these 2023PF00332 methods comprise a construction of a list of motion vector predictors (MVPs), a selection of a best MVP in the list and an encoding of the motion vector of the current block in the form of an index representing the selected MVP along, in some cases, a motion difference between the selected MVP and the actual motion vector of the current block. It appears that the representation of a motion vector in the form of an index of an MVP along with a motion vector difference carry some redundant (or residual) information on some characteristics of the motion. The existence of such residual motion information could be considered as indicating that the representation of the motion vector is sub-optimal and could still be improved. It is desirable to propose solutions allowing to overcome the above issues. In particular, it is desirable to propose solutions improving the coding of the motion information. 4. BRIEF SUMMARY In a first aspect, one or more of the present embodiments provide a method comprising: decoding an index of a first motion information predictor in an initial list of motion information predictors for a current block of a picture; decoding a motion information difference for the current block; and, applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list. 2023PF00332 In an embodiment, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. In an embodiment, the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. In an embodiment, the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. In an embodiment, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. In an embodiment, the method comprises reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference. 2023PF00332 In a second aspect, one or more of the present embodiments provide a method comprising: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor. In an embodiment, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. In an embodiment, the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of 2023PF00332 the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. In an embodiment, the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. In an embodiment, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. In an embodiment, the method comprises encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference. In a third aspect, one or more of the present embodiments provide a device comprising electronic circuitry configured for: decoding an index of a first motion information predictors in an initial list of motion information predictors for a current block of a picture; decoding a motion information difference for the current block; and, applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, 2023PF00332 adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list. In an embodiment, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. In an embodiment, the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. In an embodiment, the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. 2023PF00332 In an embodiment, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. In an embodiment, the electronic circuitry is further configured for reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference. In a fourth aspect, one or more of the present embodiments provide a device comprising electronic circuitry configured for: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor. In an embodiment, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm 2023PF00332 of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. In an embodiment, the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. In an embodiment, the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. In an embodiment, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. In an embodiment, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. In an embodiment, the electronic circuitry is further configured for encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference. In a fifth aspect, one or more of the present embodiments provide a computer program comprising program code instructions for implementing the method according to the first or the second aspect. 2023PF00332 In a sixth aspect, one or more of the present embodiments provide a non- transitory information storage medium storing program code instructions for implementing the method according to the first or the second aspect. 5. BRIEF SUMMARY OF THE DRAWINGS Fig. 1 illustrates an example of context in which various embodiments may be implemented; Fig. 2 illustrates schematically an example of partitioning undergone by a picture of pixels of an original video; Fig.3 depicts schematically a method for encoding a video stream; Fig.4 depicts schematically a method for decoding an encoded video stream; Fig. 5A illustrates schematically an example of hardware architecture of a processing module able to implement an encoding module or a decoding module in which various aspects and embodiments are implemented; Fig. 5B illustrates a block diagram of an example of a first system in which various aspects and embodiments are implemented; Fig.5C illustrates a block diagram of an example of a second system in which various aspects and embodiments are implemented; Figs.6A and 6B and illustrates schematically spatial and temporal positions considered for constructing a list of merge candidates; Figs.7A and 7B illustrates motion coding configurations in AMVP mode; Fig.8A illustrates an example of motion vector reconstruction process at a decoder side of an embodiment; Fig. 8B illustrates an example of motion vector reconstruction process at an encoder side of an embodiment; Figs.9, 11 and 13 illustrate various embodiments of an encoding algorithm; and, Figs.10, 12 and 14 illustrate various embodiments of a decoding algorithm. 6. DETAILED DESCRIPTION The following examples of embodiments are described in the context of a video format similar to VVC (ISO/IEC 23090-3 – MPEG-I : Versatile Video Coding (VVC) / ITU-T H.266). However, these embodiments are not limited to the video 2023PF00332 coding/decoding method corresponding to VVC. These embodiments are in particular adapted to various video formats comprising for example HEVC (ISO/IEC 23008-2 – MPEG-H Part 2, High Efficiency Video Coding / ITU-T H.265)), AVC ((ISO/CEI 14496-10), EVC (Essential Video Coding/MPEG-5), AV1, AV2 and VP9. Fig.1 describes an example of a context in which following embodiments can be implemented. In Fig. 1, a system 11, that could be a camera, a storage device, a computer, a server or any device capable of delivering a video stream, transmits a video stream to a system 13 using a communication channel 12. The video stream is either encoded and transmitted by the system 11 or received and/or stored by the system 11 and then transmitted. The communication channel 12 is a wired (for example Internet or Ethernet) or a wireless (for example WiFi, 3G, 4G or 5G) network link. The system 13, that could be for example a set top box, receives and decodes the video stream to generate a sequence of decoded pictures. The obtained sequence of decoded pictures is then transmitted to a display system 15 using a communication channel 14, that could be a wired or wireless network. The display system 15 then displays said pictures. In an embodiment, the system 13 is comprised in the display system 15. In that case, the system 13 and display 15 are comprised in a TV, a computer, a tablet, a smartphone, a head-mounted display, etc. Figs.2, 3 and 4 introduce an example of video format. Fig.2 illustrates an example of partitioning undergone by a picture of pixels 21 of an original video sequence 20. It is considered here that a pixel is composed of three components: a luminance component and two chrominance components. Other types of pixels are however possible comprising less or more components such as only a luminance component or an additional depth component or transparency component. A picture is divided into a plurality of coding entities. First, as represented by reference 23 in Fig. 2, a picture is divided in a grid of blocks called coding tree units (CTU). A CTU consists of an ^^ ൈ ^^ block of luminance samples together with two corresponding blocks of chrominance samples. N is generally a power of two having a maximum value of “128” for example. Second, a picture is divided into one or more groups of CTU. For example, it can be divided into one or more tile rows and tile columns, a tile being a sequence of CTU covering a rectangular region of a picture. In 2023PF00332 some cases, a tile could be divided into one or more bricks, each of which consisting of at least one row of CTU within the tile. Above the concept of tiles and bricks, another encoding entity, called slice, exists, that can contain at least one tile of a picture or at least one brick of a tile. In the example in Fig.2, as represented by reference 22, the picture 21 is divided into three slices S1, S2 and S3 of the raster-scan slice mode, each comprising a plurality of tiles (not represented), each tile comprising only one brick. As represented by reference 24 in Fig. 2, a CTU may be partitioned into the form of a hierarchical tree of one or more sub-blocks called coding units (CU). The CTU is the root (i.e. the parent node) of the hierarchical tree and can be partitioned in a plurality of CU (i.e. child nodes). Each CU becomes a leaf of the hierarchical tree if it is not further partitioned in smaller CU or becomes a parent node of smaller CU (i.e. child nodes) if it is further partitioned. In the example of Fig.2, the CTU 24 is first partitioned in “4” square CU using a quadtree type partitioning. The upper left CU is a leaf of the hierarchical tree since it is not further partitioned, i.e. it is not a parent node of any other CU. The upper right CU is further partitioned in “4” smaller square CU using again a quadtree type partitioning. The bottom right CU is vertically partitioned in “2” rectangular CU using a binary tree type partitioning. The bottom left CU is vertically partitioned in “3” rectangular CU using a ternary tree type partitioning. During the coding of a picture, the partitioning is adaptive, each CTU being partitioned so as to optimize a compression efficiency of the CTU criterion. In HEVC appeared the concept of prediction unit (PU) and transform unit (TU). Indeed, in HEVC, the coding entity that is used for prediction (i.e. a PU) and transform (i.e. a TU) can be a subdivision of a CU. For example, as represented in Fig.1, a CU of size 2 ^^ ൈ 2 ^^, can be divided in PU 2411 of size ^^ ൈ 2 ^^ or of size 2 ^^ ൈ ^^. In addition, said CU can be divided in “4” TU 2412 of size ^^ ൈ ^^ or in “16” TU of size ^^ ൈ ே ଶ ^ ^. One can note that in VVC, except in some particular cases, frontiers of the TU and PU are aligned on the frontiers of the CU. Consequently, a CU comprises generally one TU and one PU. In the present application, the term “block” or “picture block” can be used to refer to any one of a CTU, a CU, a PU and a TU. In addition, the term “block” or 2023PF00332 “picture block” can be used to refer to a macroblock, a partition and a sub-block as specified in H.264/AVC or in other video coding standards, and more generally to refer to an array of samples of numerous sizes. In the present application, the terms “reconstructed” and “decoded” may be used interchangeably, the terms “pixel” and “sample” may be used interchangeably, the terms “image,” “picture”, “sub-picture”, “slice” and “frame” may be used interchangeably. Usually, but not necessarily, the term “reconstructed” is used at the encoder side while “decoded” is used at the decoder side. Fig.3 depicts schematically a method for encoding a video stream executed by an encoding module. Variations of this method for encoding are contemplated, but the method for encoding of Fig. 3 is described below for purposes of clarity without describing all expected variations. Before being encoded, a current original picture of an original video sequence may go through a pre-processing. For example, in a step 301, a color transform is applied to the current original picture (e.g., conversion from RGB 4:4:4 to YCbCr 4:2:0), or a remapping is applied to the current original picture components in order to get a signal distribution more resilient to compression (for instance using a histogram equalization of one of the color components). Pictures obtained by pre-processing are called pre-processed pictures in the following. The encoding of a pre-processed picture begins with a partitioning of the pre- processed picture during a step 302, as described in relation to Fig.2. The pre-processed picture is thus partitioned into CTU, CU, PU, TU, etc. For each block, the encoding module determines a coding mode between an intra prediction and an inter prediction. The intra prediction consists of predicting, in accordance with an intra prediction method, during a step 303, the pixels of a current block from a prediction block derived from pixels of reconstructed blocks situated in a causal vicinity of the current block to be coded. The result of the intra prediction is a prediction mode indicating which pixels of the blocks in the vicinity to use, and a residual block resulting from a calculation of a difference between the current block and the prediction block. The inter prediction consists of predicting the pixels of a current block from a block of pixels, referred to as the reference block, of a picture preceding or following the current picture, this picture being referred to as the reference picture. During the coding of a current block in accordance with the inter prediction method, a block of the 2023PF00332 reference picture closest, in accordance with a similarity criterion, to the current block is determined by a motion estimation step 304. During step 304, a motion vector indicating the position of the reference block in the reference picture is determined. Said motion vector is used during a motion compensation step 305, involving interpolation operations between samples of the reference block, to generate a prediction block. A residual block is then calculated in the form of a difference between the current block and the prediction block. In first video compression standards, the uni-prediction inter mode described above was the only inter mode available. As video compression standards evolve, the family of inter modes has grown significantly and comprises now many different inter modes, such as bi-prediction modes in which a current is predicted from two reference blocks designated by two different motion information. During a selection step 306, the prediction mode optimising the compression performances, in accordance with a rate/distortion optimization criterion (i.e. RDO criterion), among the prediction modes tested (Intra prediction modes, Inter prediction modes), is selected by the encoding module. When the prediction mode is selected, the residual block is transformed during a step 307. In some implementations, a plurality of type of transforms can be applied to a transformed residual block. Indeed, in addition to DCT-II, a Multiple Transform Selection (MTS) scheme is used for both inter and intra predicted blocks. It uses multiple selected transforms from the DCT-VIII/DST-VII. The transformed block is then quantized during a step 309. Note that the encoding module can skip the transform and apply quantization directly to the non-transformed residual signal. The quantized residual block determined for the current block during an inter or intra prediction is encoded by an entropic encoder during a step 310. Note that the encoding module can bypass both transform and quantization, i.e., the entropic encoding is applied on the residual without the application of the transform or quantization processes. The result of the entry coding is inserted in the video data 311. When the current block is coded according to an intra prediction mode, the intra prediction mode is encoded by the entropic encoder during the step 310 in the video data 311. 2023PF00332 When the current block is encoded according to an inter prediction, a process is applied to encode the motion information. The output of this process is then encoded by the entropic encoder during the step 310 in the video data 311. Two processes are employed to encode the motion information: AMVP (Adaptive Motion Vector Prediction) or Merge. In each process, the motion information is predicted. In an implementation of the AMVP mode, a motion vector predictor (MVP) is selected, and a motion vector difference noted MVd relative to the selected MVP is computed. The MVP is selected in a list of AMVP candidates made of “2” candidates. The index of the chosen MVP and the MVd are then encoded by the entropic encoder during step 310 along with the transformed and quantized residual block resulting from the inter prediction of the current block. The AMVP candidate list is constructed first by deriving a first spatial candidate from a left block neighbouring the current block, if this block is available and inter coded. Then a second spatial candidate is derived from a top block neighbouring the current block, if this block is available and inter coded. Then, a temporal candidate is derived from a so-called collocated picture of current picture, which is an already coded picture close to current picture, at a position collocated with the current block, if an inter block exist at this collocated position. Each derived MVP candidate is scaled according to a temporal distance between the reference picture associated to this MVP candidate and the reference picture considered for the current block. A redundancy check is then conducted between derived spatial candidates and, if a duplicate candidate exists, this candidate is discarded. The final AMVP candidate list contains the two first derived MVP candidates. If less than “2” MVP candidates are obtained through the above process, then the AMVP candidate list is completed with zero motion vectors. The merge mode consists in deriving motion information of a current block from a selected motion information predictor candidate. The motion information considered here includes all the inter prediction parameters of a block, that is to say: the uni-prediction or bi-prediction type, the reference picture index within each reference picture list and the motion vector(s). The selected motion information predictor candidate (i.e., the merge candidate) is selected in a list of motion information predictor candidates (i.e., in a list of merge candidates). When a block is encoded in 2023PF00332 merge mode, the index of the selected merge candidate is encoded. If no residual block is encoded for the current block, the current block is considered as encoded according to a particular merge mode called skip mode. In some implementations, the list of merge candidates is systematically made of “5” merge candidates. Up to “5” spatial positions are considered to retrieve some potential candidates for the list of merge candidates. Fig. 6A illustrates schematically the five spatial positions considered for constructing a list of merge candidates. These positions are investigated according to the following order: 1. Left (A1) 2. Above (B1) 3. Above right (B0) 4. Left bottom (A0) 5. Above left (B2) Each spatial candidate is introduced in the list of merge candidates provided that the motion information corresponding to this candidate is not already present in the list of merge candidates. Then a temporal predictor noted TMVP is determined. Fig. 6B illustrates schematically collocated positions considered for determining the TMVP. The determination of the TMVP consists first in investigating position H and, if no motion information is available at position H, the position C is investigated. A last pruning process is then applied to ensure that the set of spatial and temporal candidates does not contain redundant candidates. In case of B-slice (slice allowing bi-predicted blocks), candidates of another type, called combined candidates, are introduced in the list of merge candidates if this list is not full. Finally, if the merge list is still not full then zero motion vectors are introduced at the end of the merge list until it is full. Recently, the representation of the motion information has slightly evolved with the apparition of two main categories of motion representation: the whole-block-based motion representation and the sub-block-based motion representation. 2023PF00332 The whole-block-based motion representation consists in assigning one set of motion information, made of one or two motion vectors and associated reference picture(s) to an inter block. Thus, the motion information of that block is represented under the form of one or two motion vectors for the whole block and a reference picture associated to each motion vector. Sub-block-based motion coding mode typically consists in dividing a block into 4x4 or 8x8 luma samples subblocks and assigning an individual set of motion information (one or two couples of a motion vector and a reference picture) to each subblock. While in previous implementations of AMVP only spatial candidates, temporal candidates and the zero-motion vector candidate were considered for constructing the list of AMVP candidates, a new category of candidates, called HMVP (History-Based Motion Vector Prediction) candidates, was added in AMVP implementations adapted to the whole block-based motion representation. A principle of HMVP candidates is to use previously coded motion vectors as MVPs. These motion vectors are associated with adjacent or non-adjacent blocks relative to a current block. To do so, a table of HMVP candidates (i.e., HMVP table) is maintained and updated on the fly, as a first-in-first-out (FIFO) buffer of MVPs. There are up to five candidates in the HMVP table. After coding one inter predicted block, provided that this block is not in sub-block mode (including affine mode) or GPM (geometric partition mode), the HMVP table is updated by appending the motion information of an inter predicted block to the end of the HMVP table as a new HMVP candidate. In addition to the usual FIFO rule, a mechanism to remove redundant HMVP candidates is applied. One can note that the HMVP table is reset at each CTU row to enable parallel processing. In recent implementations of the merge mode adapted to the whole block-based representation, the list of merge candidates was modified and three new merge modes were introduced. The list of merge candidates is constructed with the following types of candidates: ^ Spatial candidates. 2023PF00332 ^ Temporal candidates. ^ HMVP candidates. Several HMVP candidates are inserted into the list of merge candidates so that the list reaches a maximum allowed number of merge candidates minus 1. ^ Pairwise Average candidates. Up to one pairwise average candidate is added to the list of merge candidates. Pairwise candidates are computed as follows: the two first merge candidates present in the list of merge candidates are considered and their motion vectors are averaged. This averaging is computed separately for each reference picture list. If each of the two first merge candidates are bi-prediction ones, motion vectors related to both lists L0 and L1 are averaged. If only one motion vector is present, it is taken as is to form the pairwise candidate. 5. Zero motion vector candidate. The three new merge modes comprise MMVD (Merge Mode with motion vector Difference), GPM (Geometric Partitioning Mode) and CIIP (Combined Intra/Inter Prediction). These new modes are detailed in document JVET-T2002-v2: Algorithm description for Versatile Video Coding and Test Model 11 (VTM 11), Joint Video Experts Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29, 20th Meeting, by teleconference, 7 – 16 October 2020, Jianle Chen, Yan Ye, Seung Hwan Kim. MMVD can be viewed as a kind of merge mode in which a merge candidate is refined by a MVd. In MMVD, after a merge candidate is selected, it is further refined by a signalled MVd information. The signaling of a MMVD mode comprises a merge candidate flag, an index to specify a motion magnitude and an index indicating a motion direction. The merge candidate flag is signalled to specify which one is used between the first and second merge candidates. The index specifying a motion magnitude and the index indicating a motion direction allow signaling a limited number of motion vector differences (MVd) on top of a signaled merge candidate, i.e., “4” vector directions and “8” magnitude values. Fig. 7A illustrates a motion coding configuration in AMVP mode. A block to predict is shown together with his two AMVP motion vector prediction candidates^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ , and a motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ to signal. 2023PF00332 It is established that, in existing implementations of AMVP, the coding cost of a motion difference is highly correlated to the magnitude of the motion vector difference and increases as a function of the ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ magnitude. Fig.7B shows that the coding configuration of the Fig.7A can also be reached by using the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ instead of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ , leading to a smaller motion vector difference magnitude. In the following embodiments, it is shown that the situation of Fig. 7A can be detected as non-optimal by a decoding module. Indeed, in these embodiments, the decoder is able to detect that the use of motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ is not the most optimal one to employ for pointing to the spatial position corresponding to ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ + ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ , as shown in Fig. 7B. The decoder is able to detect that the use of motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ is the most likely one, compared to ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ . Therefore, it can be considered that in the configuration of Fig. 7A, the signaling of the motion vector predictor using an index of the chosen motion vector predictor and the motion vector difference carry some redundant (or residual) information that can be exploited. Metadata such as SEI (supplemental enhancement information) messages can be attached to the encoded video stream 311. A SEI message as defined for example in standards such as AVC, HEVC or VVC is a data container associated to a video stream and comprising metadata providing information relative to the video stream. After the quantization step 309, the current block is reconstructed so that the pixels corresponding to that block can be used for future predictions. This reconstruction phase is also referred to as a prediction loop. An inverse quantization is therefore applied to the transformed and quantized residual block during a step 312 and an inverse transformation is applied during a step 313. According to the prediction mode used for the block obtained during a step 314, the prediction block of the block is reconstructed. If the current block is encoded according to an inter prediction mode, the encoding module applies, when appropriate, during a step 316, a motion compensation using the motion information of the current block in order to identify each reference block of the current block. If the current block is encoded according to an intra prediction mode, during a step 315, the intra prediction mode is used for reconstructing the prediction block of the current block. The prediction block and the 2023PF00332 reconstructed residual block are added in order to obtain the reconstructed current block. Following the reconstruction, an in-loop filtering intended to reduce the encoding artefacts is applied, during a step 317, to the reconstructed block. This filtering is called in-loop filtering since this filtering occurs in the prediction loop to obtain at the decoder the same reference pictures as the encoder and thus avoid a drift between the encoding and the decoding processes. In-loop filtering tools comprises deblocking filtering, SAO (Sample adaptive Offset) and ALF (Adaptive Loop Filtering). When a block is reconstructed, it is inserted during a step 318 into a reconstructed picture stored in a memory 319 of reconstructed pictures generally called Decoded Picture Buffer (DPB). The reconstructed pictures thus stored can then serve as reference pictures for other pictures to be coded. Fig. 4 depicts schematically a method for decoding the encoded video stream 311 encoded according to method described in relation to Fig.3 executed by a decoding module. Variations of this method for decoding are contemplated, but the method for decoding of Fig. 4 is described below for purposes of clarity without describing all expected variations. The decoding is done block by block. For a current block, it starts with an entropic decoding of the current block during a step 410. Entropic decoding allows to obtain, at least, the prediction mode of the block. If the block has been encoded according to an inter prediction mode, the entropic decoding allows to obtain, when appropriate, information representative of a motion of the current block and a residual block. During a step 408, the motion information is reconstructed for the current block using the decoded information representative of the motion information. In AMVP, the information representative of the motion comprises an index of the AMVP MVP in the AMVP list and the MVd. The MVd is then added to the AMVP MVP to reconstruct the motion vector of the block. In merge, an index of the merge MVP in the merge list is obtained. The merge MVP corresponding to the index provides the motion information of the current block. In MMVD, the information representative of the motion of the current block comprises some MMVD indices representative of a MVd. The MVd corresponding to the MMVD 2023PF00332 indices is then added to the merge MVP to determine the motion information of the current block. If the block has been encoded according to an intra prediction mode, entropic decoding allows to obtain the intra prediction mode and a residual block. Steps 412, 413, 414, 415, 416 and 417 implemented by the decoding module are in all respects identical respectively to steps 412, 413, 414, 415, 416 and 417 implemented by the encoding module. Decoded blocks are saved in decoded pictures and the decoded pictures are stored in a DPB 419 in a step 418. When the decoding module decodes a given picture, the pictures stored in the DPB 419 are identical to the pictures stored in the DPB 319 by the encoding module during the encoding of said given image. The decoded picture can also be outputted by the decoding module for instance to be displayed. The post-processing step 421 can comprise an inverse color transform (e.g. conversion from YCbCr 4:2:0 to RGB 4:4:4), an inverse mapping performing the inverse of the remapping process performed in the pre-processing of step 301 and a post-filtering for improving the reconstructed pictures based for example on filter parameters provided in a SEI message. Fig. 5A illustrates schematically an example of hardware architecture of a processing module 500 able to implement an encoding module or a decoding module capable of implementing respectively a method for encoding of Fig.3 and a method for decoding of Fig. 4 modified according to different aspects and embodiments. The encoding module is for example comprised in the system 11 when this apparatus is in charge of encoding the video stream. The decoding module is for example comprised in the system 13. The processing module 500 comprises, connected by a communication bus 5005: a processor or CPU (central processing unit) 5000 encompassing one or more microprocessors, general purpose computers, special purpose computers, and processors based on a multi-core architecture, as non-limiting examples; a random access memory (RAM) 5001; a read only memory (ROM) 5002; a storage unit 5003, which can include non-volatile memory and/or volatile memory, including, but not limited to, Electrically Erasable Programmable Read-Only Memory (EEPROM), Read-Only Memory (ROM), Programmable Read-Only Memory (PROM), Random Access Memory (RAM), Dynamic Random Access Memory (DRAM), Static Random Access Memory (SRAM), flash, magnetic disk drive, and/or 2023PF00332 optical disk drive, or a storage medium reader, such as a SD (secure digital) card reader and/or a hard disc drive (HDD) and/or a network accessible storage device; at least one communication interface 5004 for exchanging data with other modules, devices or equipment. The communication interface 5004 can include, but is not limited to, a transceiver configured to transmit and to receive data over a communication channel. The communication interface 5004 can include, but is not limited to, a modem or network card. If the processing module 500 implements a decoding module, the communication interface 5004 enables for instance the processing module 500 to receive encoded video streams and to provide a sequence of decoded pictures. If the processing module 500 implements an encoding module, the communication interface 5004 enables for instance the processing module 500 to receive a sequence of original picture data to encode and to provide an encoded video stream. The processor 5000 is capable of executing instructions loaded into the RAM 5001 from the ROM 5002, from an external memory (not shown), from a storage medium, or from a communication network. When the processing module 500 is powered up, the processor 5000 is capable of reading instructions from the RAM 5001 and executing them. These instructions form a computer program causing, for example, the implementation by the processor 5000 of a decoding method as described in relation with Fig.4, an encoding method described in relation to Fig.3, and methods described in relation to Figs.6 or 7, these methods comprising various aspects and embodiments described below in this document. All or some of the algorithms and steps of the methods of Figs. 3, 4, 6 and 7 may be implemented in software form by the execution of a set of instructions by a programmable machine such as a DSP (digital signal processor) or a microcontroller, or be implemented in hardware form by a machine or a dedicated component such as a FPGA (field-programmable gate array) or an ASIC (application-specific integrated circuit). As can be seen, microprocessors, general purpose computers, special purpose computers, processors based or not on a multi-core architecture, DSP, microcontroller, FPGA and ASIC are electronic circuitry adapted to implement at least partially the methods of Figs.3, 4, 6 and 7. Fig. 5C illustrates a block diagram of an example of the system 13 in which 2023PF00332 various aspects and embodiments are implemented. The system 13 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances and head mounted display. Elements of system 13, singly or in combination, can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components. For example, in at least one embodiment, the system 13 comprises one processing module 500 that implements a decoding module. In various embodiments, the system 13 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports. In various embodiments, the system 13 is configured to implement one or more of the aspects described in this document. The input to the processing module 500 can be provided through various input modules as indicated in block 531. Such input modules include, but are not limited to, (i) a radio frequency (RF) module that receives an RF signal transmitted, for example, over the air by a broadcaster, (ii) a component (COMP) input module (or a set of COMP input modules), (iii) a Universal Serial Bus (USB) input module, and/or (iv) a High Definition Multimedia Interface (HDMI) input module. Other examples, not shown in FIG.5C, include composite video. In various embodiments, the input modules of block 531 have associated respective input processing elements as known in the art. For example, the RF module can be associated with elements suitable for (i) selecting a desired frequency (also referred to as selecting a signal, or band-limiting a signal to a band of frequencies), (ii) down-converting the selected signal, (iii) band-limiting again to a narrower band of frequencies to select (for example) a signal frequency band which can be referred to as a channel in certain embodiments, (iv) demodulating the down-converted and band- limited signal, (v) performing error correction, and (vi) demultiplexing to select the desired stream of data packets. The RF module of various embodiments includes one or more elements to perform these functions, for example, frequency selectors, signal selectors, band-limiters, channel selectors, filters, downconverters, demodulators, error correctors, and demultiplexers. The RF portion can include a tuner that performs 2023PF00332 various of these functions, including, for example, down-converting the received signal to a lower frequency (for example, an intermediate frequency or a near-baseband frequency) or to baseband. In one set-top box embodiment, the RF module and its associated input processing element receives an RF signal transmitted over a wired (for example, cable) medium, and performs frequency selection by filtering, down- converting, and filtering again to a desired frequency band. Various embodiments rearrange the order of the above-described (and other) elements, remove some of these elements, and/or add other elements performing similar or different functions. Adding elements can include inserting elements in between existing elements, such as, for example, inserting amplifiers and an analog-to-digital converter. In various embodiments, the RF module includes an antenna. Additionally, the USB and/or HDMI modules can include respective interface processors for connecting system 13 to other electronic devices across USB and/or HDMI connections. It is to be understood that various aspects of input processing, for example, Reed-Solomon error correction, can be implemented, for example, within a separate input processing IC or within the processing module 500 as necessary. Similarly, aspects of USB or HDMI interface processing can be implemented within separate interface ICs or within the processing module 500 as necessary. The demodulated, error corrected, and demultiplexed stream is provided to the processing module 500. Various elements of system 13 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards. For example, in the system 13, the processing module 500 is interconnected to other elements of said system 13 by the bus 5005. The communication interface 5004 of the processing module 500 allows the system 13 to communicate on the communication channel 12. As already mentioned above, the communication channel 12 can be implemented, for example, within a wired and/or a wireless medium. Data is streamed, or otherwise provided, to the system 13, in various embodiments, using a wireless network such as a Wi-Fi network, for example IEEE 802.11 (IEEE refers to the Institute of Electrical and Electronics Engineers). The Wi- 2023PF00332 Fi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications. The communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications. Other embodiments provide streamed data to the system 13 using the RF connection of the input block 531. As indicated above, various embodiments provide data in a non- streaming manner. Additionally, various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network. The system 13 can provide an output signal to various output devices, including the display system 15, speakers 56, and other peripheral devices 57. The display system 15 of various embodiments includes one or more of, for example, a touchscreen display, an organic light-emitting diode (OLED) display, a curved display, and/or a foldable display. The display 15 can be for a television, a tablet, a laptop, a cell phone (mobile phone), a head mounted display or other devices. The display system 15 can also be integrated with other components (for example, as in a smart phone), or separate (for example, an external monitor for a laptop). The other peripheral devices 57 include, in various examples of embodiments, one or more of a stand-alone digital video disc (or digital versatile disc) (DVR, for both terms), a disk player, a stereo system, and/or a lighting system. Various embodiments use one or more peripheral devices 57 that provide a function based on the output of the system 13. For example, a disk player performs the function of playing an output of the system 13. In various embodiments, control signals are communicated between the system 13 and the display system 15, speakers 56, or other peripheral devices 57 using signaling such as AV.Link, Consumer Electronics Control (CEC), or other communications protocols that enable device-to-device control with or without user intervention. The output devices can be communicatively coupled to system 13 via dedicated connections through respective interfaces 532, 533, and 534. Alternatively, the output devices can be connected to system 13 using the communications channel 12 via the communications interface 5004 or a dedicated communication channel via the communication interface 5004. The display system 15 and speakers 56 can be integrated in a single unit with the other components of system 13 in an electronic device such as, for example, a television. In various embodiments, the display interface 2023PF00332 532 includes a display driver, such as, for example, a timing controller (T Con) chip. The display system 15 and speaker 56 can alternatively be separate from one or more of the other components. In various embodiments in which the display system 15 and speakers 56 are external components, the output signal can be provided via dedicated output connections, including, for example, HDMI ports, USB ports, or COMP outputs. Fig. 5C illustrates a block diagram of an example of the system 11 in which various aspects and embodiments are implemented. System 11 is very similar to system 13. The system 11 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, a camera and a server. Elements of system 11, singly or in combination, can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components. For example, in at least one embodiment, the system 11 comprises one processing module 500 that implements an encoding module. In various embodiments, the system 11 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports. In various embodiments, the system 11 is configured to implement one or more of the aspects described in this document. The input to the processing module 500 can be provided through various input modules as indicated in block 531 already described in relation to Fig.5C. Various elements of system 11 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards. For example, in the system 11, the processing module 500 is interconnected to other elements of said system 11 by the bus 5005. The communication interface 5004 of the processing module 500 allows the system 500 to communicate on the communication channel 12. Data is streamed, or otherwise provided, to the system 11, in various embodiments, using a wireless network such as a Wi-Fi network, for example IEEE 802.11 (IEEE refers to the Institute of Electrical and Electronics Engineers). The Wi- 2023PF00332 Fi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications. The communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications. Other embodiments provide streamed data to the system 11 using the RF connection of the input block 531. As indicated above, various embodiments provide data in a non-streaming manner. Additionally, various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network. The data provided to the system 11 can be provided in different format. In various embodiments these data are encoded and compliant with a known video compression format such as AV1, VP9, VVC, HEVC, AVC, etc. In various embodiments, these data are raw data provided for example by a picture and/or audio acquisition module connected to the system 11 or comprised in the system 11. In that case, the processing module take in charge the encoding of these data. The system 11 can provide an output signal to various output devices capable of storing and/or decoding the output signal such as the system 13. Various implementations involve decoding. “Decoding”, as used in this application, can encompass all or part of the processes performed, for example, on a received encoded video stream in order to produce a final output suitable for display. In various embodiments, such processes include one or more of the processes typically performed by a decoder, for example, entropy decoding, inverse quantization, inverse transformation, and prediction. In various embodiments, such processes also, or alternatively, include processes performed by a decoder of various implementations described in this application, for example, for reconstruction motion information. Whether the phrase “decoding process” is intended to refer specifically to a subset of operations or generally to the broader decoding process will be clear based on the context of the specific descriptions and is believed to be well understood by those skilled in the art. Various implementations involve encoding. In an analogous way to the above discussion about “decoding”, “encoding” as used in this application can encompass all or part of the processes performed, for example, on an input video sequence in order to 2023PF00332 produce an encoded video stream. In various embodiments, such processes include one or more of the processes typically performed by an encoder, for example, partitioning, prediction, transformation, quantization, and entropy encoding. In various embodiments, such processes also, or alternatively, include processes performed by an encoder of various implementations described in this application, for example, for predicting motion information. Whether the phrase “encoding process” is intended to refer specifically to a subset of operations or generally to the broader encoding process will be clear based on the context of the specific descriptions and is believed to be well understood by those skilled in the art. Note that the syntax elements names as used herein, are descriptive terms. As such, they do not preclude the use of other syntax element names. When a figure is presented as a flow diagram, it should be understood that it also provides a block diagram of a corresponding apparatus. Similarly, when a figure is presented as a block diagram, it should be understood that it also provides a flow diagram of a corresponding method/process. Various embodiments refer to rate distortion optimization. In particular, during the encoding process, the balance or trade-off between a rate and a distortion is usually considered. The rate distortion optimization is usually formulated as minimizing a rate distortion function, which is a weighted sum of the rate and of the distortion. There are different approaches to solve the rate distortion optimization problem. For example, the approaches may be based on an extensive testing of all encoding options, including all considered modes or coding parameters values, with a complete evaluation of their coding cost and related distortion of a reconstructed signal after coding and decoding. Faster approaches may also be used, to save encoding complexity, in particular with computation of an approximated distortion based on a prediction or a prediction residual signal, not the reconstructed one. Mix of these two approaches can also be used, such as by using an approximated distortion for only some of the possible encoding options, and a complete distortion for other encoding options. Other approaches only evaluate a subset of the possible encoding options. More generally, many approaches employ any of a variety of techniques to perform the optimization, but the optimization is not necessarily a complete evaluation of both the coding cost and related distortion. 2023PF00332 The implementations and aspects described herein can be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed can also be implemented in other forms (for example, an apparatus or program). An apparatus can be implemented in, for example, appropriate hardware, software, and firmware. The methods can be implemented, for example, in a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs"), and other devices that facilitate communication of information between end-users. Reference to “one embodiment” or “an embodiment” or “one implementation” or “an implementation”, as well as other variations thereof, means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment. Thus, the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout this application are not necessarily all referring to the same embodiment. Additionally, this application may refer to “determining” various pieces of information. Determining the information can include one or more of, for example, estimating the information, calculating the information, predicting the information, retrieving the information from memory or obtaining the information for example from another device, module or from user. Further, this application may refer to “accessing” various pieces of information. Accessing the information can include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, moving the information, copying the information, calculating the information, determining the information, predicting the information, or estimating the information. Additionally, this application may refer to “receiving” various pieces of information. Receiving is, as with “accessing”, intended to be a broad term. Receiving the information can include one or more of, for example, accessing the information, or 2023PF00332 retrieving the information (for example, from memory). Further, “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information. It is to be appreciated that the use of any of the following “/”, “and/or”, and “at least one of”, “one or more of” for example, in the cases of
Figure imgf000031_0001
“A and/or B” and “at least one of A and B”, “one or more of A and B” is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of both options (A and B). As a further example, in the cases of “A, B, and/or C” and “at least one of A, B, and C”, “one or more of A, B and C” such phrasing is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of the third listed option (C) only, or the selection of the first and the second listed options (A and B) only, or the selection of the first and third listed options (A and C) only, or the selection of the second and third listed options (B and C) only, or the selection of all three options (A and B and C). This may be extended, as is clear to one of ordinary skill in this and related arts, for as many items as are listed. Also, as used herein, the word “signal” refers to, among other things, indicating something to a corresponding decoder. For example, in certain embodiments the encoder signals a use of some coding tools. In this way, in an embodiment the same parameters can be used at both the encoder side and the decoder side. Thus, for example, an encoder can transmit (explicit signaling) a particular parameter to the decoder so that the decoder can use the same particular parameter. Conversely, if the decoder already has the particular parameter as well as others, then signaling can be used without transmitting (implicit signaling) to simply allow the decoder to know and select the particular parameter. By avoiding transmission of any actual functions, a bit savings is realized in various embodiments. It is to be appreciated that signaling can be accomplished in a variety of ways. For example, one or more syntax elements, flags, and so forth are used to signal information to a corresponding decoder in various embodiments. While the preceding relates to the verb form of the word “signal”, the word “signal” can also be used herein as a noun. 2023PF00332 As will be evident to one of ordinary skill in the art, implementations can produce a variety of signals formatted to carry information that can be, for example, stored or transmitted. The information can include, for example, instructions for performing a method, or data produced by one of the described implementations. For example, a signal can be formatted to carry the encoded video stream and SEI messages of a described embodiment. Such a signal can be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal. The formatting can include, for example, encoding an encoded video stream and modulating a carrier with the encoded video stream. The information that the signal carries can be, for example, analog or digital information. The signal can be transmitted over a variety of different wired or wireless links, as is known. The signal can be stored on a processor-readable medium. The following embodiments try to take advantage of the redundant (or residual) information existing in the encoding of the motion information in the form of an index on a MVP and a MVd. Fig. 8A illustrates an example of motion vector reconstruction process at decoder side of an embodiment. Suppose a motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ as in Fig. 8A is signaled in video data along with a syntax element mvp_flag indicating a use of motion vector ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ as the motion vector predictor of a current block. In this embodiment, a decoding-side validity checking process is applied to determine a validity of the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ (represented by the index mvp_flag) and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ . The validity checking process is based on the following decoding-side criterion: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ .൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ൯ ^ ଶ ฮ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ଶ ^ ฮ (Eq. 1) As can be vector
Figure imgf000032_0001
difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ and the motion information predictors of the AMVP list (i.e., ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^. This decoding side criterion is easy to compute and is strictly equivalent to the following second decoding side criterion: 2023PF00332 ฮ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ ฮ ^ ฮ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ଶ ^
Figure imgf000033_0001
Geometrically, the decoding side validity checking process checks if the motion vector reconstructed with ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ points to a spatial position that can be reached with a motion vector difference smaller than ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ in magnitude, and from motion vector ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ . If this is the case, in this embodiment, the AMVP list is adapted to obtain an adapted list. The adaptation of the AMVP list consists in replacing the motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ by an alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ , so that the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ represents the motion vector that is actually reconstructed by the decoder. Spatial position ^^^ shown by Fig.8A. corresponds to the spatial position pointed to by the adaptive motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ . Intuitively, from the illustration of Fig. 8A, one understands that the motion vector difference required to reach the target spatial position ( ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ ^ starting from spatial position ^^^ is of smaller magnitude than the MVD required when starting from spatial position ^^^, hence leads to lower MVD coding rate cost. The alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ is obtained as follows: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ 1 ^ ^ 2൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯ The combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is represented by a final motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^: ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ ൌ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ On Fig.8A, the spatial region ^^^ is the set of spatial positions that are reached from spatial position ^^^ with a smaller MVD than from spatial position ^^^, where ^^^ and ^^^ are the spatial positions pointed to respectively by ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ . The region ^^^ is the set of spatial positions that are reached from spatial position ^^^ with a smaller MVD than from spatial position ^^^. Finally, the proposed embodiment leads to the establishment of a further spatial region ^^^ , shown on Fig.8A. Region ^^^ correspond to spatial positions that can be reached from position ^^^ with a smaller MVD in magnitude than from both positions ^^^and ^^^. As a consequence, thanks to the proposed embodiment, coding a motion vector pointing to a position in ^^^ is achieved with 2023PF00332 smaller MVD magnitudes than in prior methods, hence reduced MVD coding cost. Thus, increased compression is expected from the proposed embodiment. The decoding-side validity checking process is used together with a compliant encoding-side validity checking process illustrated by Fig.8B. Assume a motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ is to be encoded in video data for a current block. Motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ points to a spatial position ^^^^^^^ as shown in Fig. 8B. Position ^^^^^^^ is closer to position ^^^ pointed to by ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ than position ^^^ pointed to by ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ . In a usual process, the motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ is selected to predict the motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ and a motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ is computed between ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^. An encoding side validity checking process checks if the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ and the motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ can be modified to produce a smaller motion vector difference magnitude, and in a way detectable by a decoder. The encoding-side validity checking process comprises checking the following encoding side criterion: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ . ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^^^^^ ^^^^^^^^^^^ ^^^^^^^^^^^ ଶ ൫ ^ ^ ^^^ ൯ ^ ฮ ^^ ^^ ^^^ െ ^^ ^^ ^^^ ฮ ... (Eq. 2) This if the
Figure imgf000034_0001
combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ form a valid combination. If the encoding side criterion is verified, i.e., if the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is invalid, then the encoding module replaces the motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ by an alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ : ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ 1 ^ ^ 2 ൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯ In addition, the motion vector difference to encode is modified to the following value: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ This
Figure imgf000034_0002
The following proves this alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ and the 2023PF00332 recomputed motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ reduces the motion vector difference magnitude to code: We have: ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ ฮ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ .൫ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ 1 ଶ ^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯ ^ 4 ฮ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ฮ
Figure imgf000035_0004
^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ ฮ െ ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ଶ 1 ^ ฮ ^ 4 ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ଶ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ฮ
Figure imgf000035_0001
^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ
Figure imgf000035_0002
The following proves the alternative motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ and the recomputed motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ leads to a configuration that is detectable by the decoder: We have: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ^ .൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯ ൌ ൬ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ െ ଶ ൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯^
Figure imgf000035_0003
Therefore, the motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ fulfills the decoding-side criterion (Equation Eq. 1) if the motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ is signaled together with the use of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ as the motion vector predictor of the considered block. It is then possible to signal a motion vector difference with reduced magnitude compared to what would be signaled in existing implementations, while making it possible to reconstruct the target final motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ on the decoder side. 2023PF00332 Note the encoder side criterion of equation Eq. 2 can be generalized to the following expression: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ .൫ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ଶ ^ ൯ ^ ^ଶ ^ ^^^ . ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ฮ ...(Eq. 3)
Figure imgf000036_0001
Where ^^ is a strictly positive value. Note in above exemplary embodiment, ^^ has value ^. In that case, the alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ is computed as follows: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ ^^.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯ The motion
Figure imgf000036_0002
^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ᇱ^^ ^ ൌ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ െ ^^.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ൯
Figure imgf000036_0003
Fig.9 illustrates an embodiment of an encoding algorithm. The encoding algorithm is for example executed by the processing module 500 of the system 11 when the system 11 implements the encoding module of Fig. 3. The encoding algorithm is for example executed during step 308. In a step 901, the processing module 500 obtains a motion vector predictor for a motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ of a current block (obtained in step 304). During step 901, the processing module 500 selects the motion vector predictor minimizing the motion vector difference among the motion vector predictor candidates ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ of the AMVP list. The selected motion vector predictor candidate, noted ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ , is represented by its index in the AMVP list idx. In a step 902, the processing module 500 the motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ : ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ ൌ ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ௗ௫ In step 903 and 904, the processing module 500 applies a validity checking process. The validity checking process comprises determining from the encoding side criterion of equation Eq. 3 that the non-selected motion vector predictor of the AMVP list ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference form a valid combination: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ .൫ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ^ ଶ ି^^ ^^ ^^ ௫൯ ^ ^ଶ ^ ^^^ . ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫ ฮ (Eq. 4)
Figure imgf000036_0004
In step 903, the processing module 500 computes the encoding side criterion. In step 904, the processing module 500 determines if the encoding side criterion 2023PF00332 is verified. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is valid (i.e., if the condition of the encoding side criterion is not verified), step 904 is followed by a step 909. In step 909, the processing module 500 encodes the motion vector predictor index idx in the video data 311. In a step 910, the processing module 500 encodes the motion vector difference^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ in the video data 311. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is not valid (i.e., if the condition of the encoding side criterion is fulfilled), step 904 is followed by a step 905. During step 905, the processing module 500 adapt the AMVP list to obtain an adapted list. The adapting of the AMVP list comprises replacing the motion vector predictor candidate ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ by an alternative motion vector predictor ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ௫ : ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ௫ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^^ ௫ ^ ^^.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫൯ In a the motion vector
Figure imgf000037_0001
difference based on the adapted list. The recomputed motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ is computed as follows: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ െ ^^.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫൯ In a step motion vector predictor
Figure imgf000037_0002
index 1-idx in the video data 311. In a step 908, the processing module 500 encodes the motion vector difference^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ in the video data 311. Fig.10 illustrates an embodiment of a decoding algorithm. The decoding algorithm is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of Fig. 4. The decoding algorithm is for example executed during step 408. In a step 1001, the processing module decodes an index ^^ ^^ ^^^^^ of a motion vector predictor for a current block from the video data 311. This index ^^ ^^ ^^^^^ identifies a motion vector predictor among the motion vector predictor candidates^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ of the AMVP list. 2023PF00332 In a step 1002, the processing module 500 decodes a motion vector difference ^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^^^^^^^^^ for the current block from the video data 311. In steps 1003 and 1004, the processing module 500 applies a decoding-side validity checking process. The decoding-side validity checking process comprises determining from the decoding-side criterion of equation Eq. 1 that the motion information predictor^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ indicated by the index ^^ ^^ ^^ ^^^ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^^^^^^^^^ form a valid combination. In step 1003, the processing module 500 computes the decoding side criterion: ^ ^^^ ^ ^^ ^ ^ ^ ^^^^^^^^^^^ .൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ವ^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ 1 ^ௗ௫ವ^^൯ ^ 2 ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ଶ ^^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ ฮ
Figure imgf000038_0001
In step 1004, the processing module 500 determines if the combination is valid, i.e. if the above decoding-side criterion is not fulfilled. If the combination is valid, in a step 1007, the processing module 500 reconstructs the motion vector of the current block ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ as the sum between the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ and the motion vector difference ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^^^^^^^^^ : ^ ^^^^^^ ^^^^^^^^ ൌ ^^^^^^^ ^^^^^ ^^^^^ప^^ௗ^^௫^^ವ^^^^^^ ^ ^ ^^^ ^ ^^ ^ ^ ^ ^^ ^^^^^^^^^
Figure imgf000038_0002
If the combination is not valid (i.e. if the decoding-side criterion is fulfilled), in a step 1005, the processing module 500 adapt the AMVP list to obtain an adapted list. The adaptation of the AMVP list comprises replacing the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ by an alternative motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ : ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^^ ^^ ^^^ ^ ^ ^ ൌ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ ^ ^^.൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ െ ^ ^^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ ൯ In vector of the
Figure imgf000038_0003
current block ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ as the sum between the motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ and the motion vector difference ^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^^^^^^^^^ : ^ ^^^^^^ ^^^^^^^^ ൌ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ ^ ^ ^^^ ^ ^^ ^ ^ ^ ^^ ^^^^^^^^^ In a first variant, more variety is given for choosing the alternative motion 2023PF00332 vector predictor by offering the possibility of using the zero-motion-vector. To do so, when the decoding-side criterion indicates that the combination of ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^^^^^^^^^ is not valid, an additional decoding-side criterion is computed during step 1005 to determine if the alternative motion vector predictor is the one computed in step 1005 (i.e., ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^^ ^^ ^^^ ^ ^ ^ ) or is the zero-motion- vector. For example, the additional decoding-side criterion consists in checking the following condition: ฮ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ ฮ ^ ^^. min ^ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ଶ ^ ฮ , ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ଶ ^ ฮ ^ where 0 ^ ^^ a and b.
Figure imgf000039_0001
Responsive to decoding-side criterion and the second decoding side criterion are both fulfilled, then, during step 1005, the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ is replaced by the zero-motion vector. In this first variant, when the decoding-side criterion is applied by the processing module 500 of the system 13, an additional encoding-side criterion identical to the additional decoding-side criterion is applied by the processing module 500 of the system 11 in step 905 to determine if the alternative motion vector predictor is the one computed in step 905 (i.e., ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ௫ ) or is the zero-motion-vector. In a second variant, the decoding-side criterion of equation Eq.1 is used by the encoding module and the decoding module. To avoid confusion, the decoding-side criterion of equation Eq.1 is called validity criterion in the following and is used in its generalized form represented by equation Eq.5 below: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ .൫ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫൯ ^ ^ ^^ . ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ଶ ฮ (Eq. 5)
Figure imgf000039_0002
When the validity criterion of equation Eq. 5 is fulfilled, the encoding and the decoding modules consider a next candidate in the succession of potential candidates considered during the AMVP and/or merge candidates list construction process. For instance, w.r.t. AMVP, spatial candidates, a temporal motion vector candidate, HMVP candidates and then the zero-motion candidate may be considered during the AMVP candidate list construction process. In the second variant, potential candidates considered in the process for constructing a list of motion vector predictors are used to 2023PF00332 generate a further motion vector predictor, noted ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ , different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ . The second variant allows to diversify the set of potential motion vector predictors that can be used in combination with a given motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ . Fig.11 illustrates an embodiment of an encoding algorithm compliant with the second variant. The encoding algorithm of Fig. 11 is for example executed by the processing module 500 of the system 11 when the system 11 implements the encoding module of Fig.3. The encoding algorithm of Fig.11 is for example executed during step 308. Steps 901, 902, 909 and 910 remains unchanged compared to the algorithm of Fig.9. In steps 1100 and 1101, the processing module 500 applies a validity checking process. The validity checking process comprises determining from the validity criterion of equation Eq. 1 that the non-selected motion vector predictor of the AMVP list ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ form a valid combination: ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^ ^ ^^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^ ଶ .൫ ^^ ^^ ^^పௗ௫ െ ^^ ^^ ^^^ିపௗ௫൯ ^ ^ଶ^ . ฮ ^^ ^^ ^^పௗ௫ െ ^^ ^^ ^^^ିపௗ௫ ฮ criterion.
Figure imgf000040_0001
In step 1101, the processing module 500 determines if the validity criterion is verified. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is not valid (i.e., if the condition of the validity criterion is fulfilled), step 1101 is followed by a step 1102. In the step 1102, the processing module 500 tries to obtain a next candidate^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. For instance, the next candidate ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ is the first candidate from the potential candidates considered during the AMVP and/or merge candidates list construction process different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ . In a step 1103, the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 909. Otherwise, step 1103 is followed by a step 1105. During step 1105, the processing module 500 recomputes a motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ between the 2023PF00332 motion vector ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ of a current block and the next candidate ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ ൌ ^ ^ ^^ ^ ^^ ^ ^ ^^^^^^ െ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ In a step 1106, the the validity criterion on the
Figure imgf000041_0001
combination of the motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ and the non-selected motion vector predictor of the AMVP list ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ : ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ .൫ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ௫൯ ^ ^ ^^ . ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^^ ^^ ௫ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ଶ ฮ (Eq. 6)
Figure imgf000041_0002
As can be seen, Equation Eq. 6 correspond to equation Eq. 5 adapted to the combination of the motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ and the non-selected motion vector predictor of the AMVP list ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫. In a step 1107, the processing module 500 determines if the validity criterion is verified. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ is valid (i.e., if the condition of the above criterion of equation Eq.5 is not verified), step 1107 is followed by a step 909. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ is not valid, step 1107 is followed by step 1108. In step 1108, the processing module 500 compares the magnitude of the motion vector difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ௗ to the magnitude of the motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ . If ^ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^^ ଶ ^ ^ ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ଶ ฮ , step 1108 is followed by step 909. If ^ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^^ ଶ ^ ^ ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ ฮ , step 1108 is followed by steps 907 and 908 already explained in relation to Fig.9. Fig. 12 illustrates an embodiment of a decoding algorithm compliant with the second variant. The encoding algorithm of Fig. 12 is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of Fig.4. The encoding algorithm of Fig.13 is for example executed during step 408. Steps 1001, 1002, 1003, 1004 and 1007 remains unchanged compared to the algorithm of Fig.10. If the combination is not valid (i.e. if the validity criterion of equation Eq. 5 is 2023PF00332 fulfilled), step 1004 is followed by a step 1201. During step 1201, the processing module 500 tries to obtain a next candidate^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. Again, for instance, the next candidate ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ is the first candidate from the potential candidates considered during the AMVP and/or merge candidates list construction process different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ . In a step 1202, the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 1007. If a next candidate can be obtained, the processing module 500 applies a step 1204. During step 1204 the processing module 500 adapts the AMVP list to obtain an adapted list. The adaptation of the AMVP list comprises replacing the motion vector predictor ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ by an alternative motion vector predictor ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ^^ ^^ ^^^ ^ ^ ^ : ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^^^ ^^ ^^^ ^ ^ ^ ൌ ^ ^^^^^^ ^^^^^ ^^^^^^^^^^^௫^^௧ Step 1204 is followed by step 1006, already explained in relation to Fig.10. In a third variant, the validity criterion is used by the encoding module and the decoding module. When the validity criterion is fulfilled, the encoding and the decoding modules evaluate potential candidates considered during the AMVP and/or merge candidates list construction process (for instance, w.r.t. AMVP, spatial candidates, a temporal motion vector candidate, HMVP candidates and then the zero-motion) until finding a candidate sufficiently different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ௫. In the third variant, if a candidate sufficiently different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫ is identified, this identified candidate is used to generate a further motion vector predictor, noted ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ , different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ and ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ . The third variant allows to increase the diversity of potential motion vector predictors that can be used in combination with a given motion vector difference ^ ^^^ ^ ^^ ^ ^^ ^ ^ ௗ . Fig.13 illustrates an embodiment of an encoding algorithm compliant with the third variant. The encoding algorithm of Fig. 13 is for example executed by the processing 2023PF00332 module 500 of the system 11 when the system 11 implements the encoding module of Fig.3. The encoding algorithm of Fig.13 is for example executed during step 308. Steps 901, 902, 1100, 1101, 909 and 910 remains unchanged compared to the same steps in the algorithms of Figs.9 and 11. If the combination of ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^ି^ௗ௫ and the motion information difference ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ௗ is not valid (i.e., if the condition of validity criterion of Equation Eq. 5 is fulfilled), step 1101 is followed by an iterative process comprising steps 1102, 1103 and a new step 1104 allowing evaluating the potential candidates considered during the AMVP and/or merge candidates list construction process. Potential candidates are successively considered (i.e. run though) in the order in which they are considered for the construction of the AMVP and/or the merge list. In the step 1102, the processing module 500 tries to obtain a next candidate^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. In the step 1103, the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 909. Otherwise, step 1103 is followed by step 1104. During step 1104, the processing module 500 determines if the obtained candidate is sufficiently different from^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ௫. To do so, the processing module 500 compares, for example, the following metric to a value Th: ฮ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫ ฮ ^ ^^ℎ If ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ௫ ฮ ^ Th, the processing module 500 loops back to step 1102 to evaluate a new next candidate ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. If ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^ ^ ^^ ି ^^^ ^^ ^^ ௫ ฮ ^ ^^ℎ, step 1104 is followed by steps 1105, 1106, as already explained in relation to Fig.11.
Figure imgf000043_0001
Fig. 14 illustrates an embodiment of a decoding algorithm compliant with the third variant. The encoding algorithm of Fig. 14 is for example executed by the processing module 500 of the system 13 when the system 13 implements the decoding module of 2023PF00332 Fig.4. The encoding algorithm of Fig.13 is for example executed during step 408. Again, steps 1001, 1002, 1003, 1004 and 1007 remains unchanged compared to the algorithm of Fig.10. If the combination is not valid (i.e., if the validity criterion of equation Eq.5 is fulfilled), step 1004 is followed by an iterative process comprising steps 1201, 1202 and a new step 1203 allowing evaluating the potential candidates considered during the AMVP and/or merge candidates list construction process. Potential candidates are run though in the order in which they are considered for the construction of the AMVP and/or the merge list. During step 1201, the processing module 500 tries to obtain a next candidate^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. In a step 1202, the processing module 500 determines if a next candidate can be obtained. If no next candidate can be obtained, the processing module 500 continues with step 1007. If a next candidate can be obtained, the processing module 500 applies a step 1203. During step 1203, the processing module 500 determines if the obtained candidate is sufficiently different from ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ . To do so, the processing module 500 compares, for example, the following metric to a value Th: ฮ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ ฮ ^ ^^ℎ If ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ ฮ ^ Th, the processing module 500 loops back to step 1201 to search a new next candidate ^ ^^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^^ ^^ ௧ from the potential candidates considered during the AMVP and/or merge candidates list construction process. If ฮ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^ ^^ െ ^ ^ ^^ ^ ^^ ^ ^^ ^ ^ ^ ^^ ^ ^ ^ ^^^ ି ^^ ^^ ^^ ^^ ^^^ ^ ^ ^ ฮ ^ ^^ℎ, step 1203 is followed by steps 1204 and 1006 as already Fig.12.
Figure imgf000044_0001
We described above a number of embodiments. Features of these embodiments can be provided alone or in any combination. Further, embodiments can include one or more of the following features, devices, or aspects, alone or in any combination, across various claim categories and types: 2023PF00332 ^ A bitstream or signal that includes one or more of the described syntax elements, or variations thereof. ^ Creating and/or transmitting and/or receiving and/or decoding a bitstream or signal that includes one or more of the described syntax elements, or variations thereof. ^ A TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described. ^ A TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described, and that displays (e.g. using a monitor, screen, or other type of display) a resulting picture. ^ A TV, set-top box, cell phone, tablet, or other electronic device that tunes (e.g. using a tuner) a channel to receive a signal including an encoded video stream, and performs at least one of the embodiments described. ^ A TV, set-top box, cell phone, tablet, or other electronic device that receives (e.g. using an antenna) a signal over the air that includes an encoded video stream, and performs at least one of the embodiments described. ^ A server, camera, cell phone, tablet or other electronic device that transmits (e.g. using an antenna) a signal over the air that includes an encoded video stream, and performs at least one of the embodiments described. ^ A server, camera, cell phone, tablet or other electronic device that tunes (e.g. using a tuner) a channel to transmit a signal including an encoded video stream, and performs at least one of the embodiments described.

Claims

2023PF00332 Claims 1. A method comprising: decoding (1001) an index of a first motion information predictor in an initial list of motion information predictors for a current block of a picture; decoding (1002) a motion information difference for the current block; and, applying (1003, 1004) a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, adapting (1005) the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. 2. The method of claim 1 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list. 3. The method of claim 1 or 2 wherein, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. 4. The method of claim 3 wherein the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. 2023PF00332 5. The method of claim 1 or 2 wherein the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. 6. The method of claim 5 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. 7. The method of claim 6 wherein, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. 8. The method of any previous claim comprising reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference. 9. A method comprising: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. 2023PF00332 10. The method of claim 9 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor. 11. The method of claim 9 or 10 wherein, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. 12. The method of claim 11 wherein the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. 13. The method of claim 9 or 10 wherein the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. 14. The method of claim 13 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. 2023PF00332 15. The method of claim 14 wherein, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. 16. The method of any previous claim from claim 10 to 15 comprising encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference. 17. A device comprising electronic circuitry configured for: decoding (1001) an index of a first motion information predictor in an initial list of motion information predictors for a current block of a picture; decoding (1002) a motion information difference for the current block; and, applying (1003, 1004) a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that the first motion information predictor and the motion information difference form a valid combination; and, adapting (1005) the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. 18. The device of claim 17 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list. 19. The device of claim 17 or 18 wherein, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of the lowest norm of norms of the motion information predictors of the initial list and selecting the alternative motion information predictor among a motion 2023PF00332 information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. 20. The device of claim 19 wherein the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the a value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. 21. The device of claim 17 or 18 wherein the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. 22. The device of claim 21 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. 23. The device of claim 22 wherein, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. 24. The device of any previous claim from claim 17 to 23 comprising reconstructing the motion information of the current block as a sum of the alternative motion information predictor and the motion information difference. 25. A device comprising electronic circuitry configured for: obtaining a first motion information predictor of an initial list of motion information predictors for a current block of a picture; computing a motion information difference between a motion information of the current block and the first motion information predictor; 2023PF00332 applying a validity checking process comprising determining from a criterion calculated from the motion information difference and the motion information predictors of the initial list that a second motion information predictor of the initial list and the motion information difference form a valid combination; and, adapting the initial list of motion information predictors to obtain an adapted list based on a result of the applying of the validity checking process. 26. The device of claim 25 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor, the alternative motion information predictor being determined based on the motion information predictors of the initial list; and recomputing the motion information difference based on the adapted list comprises recomputing the motion information difference as a difference between the motion information of the current block and the alternative motion information predictor. 27. The device of claim 25 or 26 wherein, determining the alternative motion information predictor based on the motion information predictors of the initial list comprises comparing a norm of the motion information difference with a value representative of a lowest norm of norms of the motion information predictors of the initial list, selecting the alternative motion information predictor among a motion information predictor derived from the motion information predictors of the initial list and a zero motion vector based on a result of the comparison. 28. The device of claim 27 wherein the zero motion vector is selected responsive to a ratio between the norm of the motion information difference and the value representative of the lowest norm of the norms of the motion information predictors of the initial list is lower than a value. 29. The device of claim 25 or 26 wherein the initial list of motion information predictors comprises the first motion information predictor and a second motion information predictor, and the criterion is based on the first motion information predictor and the second motion information predictor. 2023PF00332 30. The device of claim 29 wherein, responsive to the first motion information predictor and the motion information difference form a non-valid combination, adapting the initial list comprises replacing the first motion information predictor by an alternative motion information predictor corresponding to a candidate selected among potential candidates considered during an AMVP and/or merge candidates list construction process. 31. The device of claim 30 wherein, the selected candidate is such that a value representative of a difference between the selected candidate and the second motion information predictor is higher than a value. 32. The device of any previous claim from claim 26 to 31 comprising encoding the motion information of the current block in the form of an index representative of the first motion information predictor and the recomputed motion information difference. 33. A computer program comprising program code instructions for implementing the method according to any previous claims from claim 1 to 16. 34. Non-transitory information storage medium storing program code instructions for implementing the method according to any previous claims from claim 1 to 16.
PCT/EP2024/061519 2023-05-15 2024-04-26 Removing some redundancies in motion information coding Pending WO2024235608A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
EP23305765 2023-05-15
EP23305765.2 2023-05-15
EP24305047 2024-01-09
EP24305047.3 2024-01-09

Publications (1)

Publication Number Publication Date
WO2024235608A1 true WO2024235608A1 (en) 2024-11-21

Family

ID=90904307

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2024/061519 Pending WO2024235608A1 (en) 2023-05-15 2024-04-26 Removing some redundancies in motion information coding

Country Status (1)

Country Link
WO (1) WO2024235608A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012090397A1 (en) * 2010-12-28 2012-07-05 株式会社Jvcケンウッド Video encoding device, video encoding method, and video encoding program, and video decoding device, video decoding method, and video decoding program
KR20180121862A (en) * 2010-01-19 2018-11-09 삼성전자주식회사 Method and apparatus for encoding/decoding motion vector based on reduced motion vector predictor candidates
US20220070456A1 (en) * 2018-12-21 2022-03-03 Mediatek Inc. Method and Apparatus of Simplified Triangle Merge Mode Candidate List Derivation

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20180121862A (en) * 2010-01-19 2018-11-09 삼성전자주식회사 Method and apparatus for encoding/decoding motion vector based on reduced motion vector predictor candidates
WO2012090397A1 (en) * 2010-12-28 2012-07-05 株式会社Jvcケンウッド Video encoding device, video encoding method, and video encoding program, and video decoding device, video decoding method, and video decoding program
US20220070456A1 (en) * 2018-12-21 2022-03-03 Mediatek Inc. Method and Apparatus of Simplified Triangle Merge Mode Candidate List Derivation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JIANLE CHENYAN YESEUNG HWAN KIM: "JVET-T2002-v2: Algorithm description for Versatile Video Coding and Test Model 11 (VTM 11)", JOINT VIDEO EXPERTS TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29, 20TH MEETING, 7 October 2020 (2020-10-07)

Similar Documents

Publication Publication Date Title
US20220159277A1 (en) Method and apparatus for video encoding and decoding with subblock based local illumination compensation
EP4289141A1 (en) Spatial local illumination compensation
US12041241B2 (en) Method and device for image encoding and decoding with a shifted position
KR20210062055A (en) Method and apparatus for video encoding and decoding using bidirectional prediction
US20250024067A1 (en) Video encoding and decoding using reference picture resampling
US12335529B2 (en) Method and device for consistent interaction of signaling tools with encoding tools in video decoding
US12143567B2 (en) Subblock merge candidates in triangle merge mode
US20250030834A1 (en) Method and device for picture encoding and decoding
EP3854078A1 (en) Translational and affine candidates in a unified list
WO2024002675A1 (en) Simplification for cross-component intra prediction
US20240171731A1 (en) Geometric partitions with switchable interpolation filter
WO2020112451A1 (en) Combining affine candidates
WO2024235608A1 (en) Removing some redundancies in motion information coding
EP3991428A1 (en) Hmvc for affine and sbtmvp motion vector prediction modes
WO2020193736A1 (en) Inter-prediction parameter derivation for video encoding and decoding
EP4629615A1 (en) Combining the intra sub-partition (isp) mode and the combination intra-inter partition mode (ciip)
WO2024235610A1 (en) Inferring motion vector predictor based on motion vector difference
EP4622255A1 (en) Template based intra prediction with mode split
WO2024223400A1 (en) Motion information predictor selection
WO2025002767A1 (en) New hypothesis for multi-hypothesis inter prediction mode
WO2025149461A1 (en) Method and apparatus for encoding/decoding with geometric partitioning mode
EP4505716A1 (en) Temporal intra mode derivation
WO2025067883A1 (en) Deriving a coding mode from one or more template-based costs
EP4602799A1 (en) Template type selection for video coding and decoding
WO2025149307A1 (en) Combination of extrapolation filter-based intra prediction with other intra prediction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 24722032

Country of ref document: EP

Kind code of ref document: A1