Search

EP-4736452-A1 - METHOD, APPARATUS, AND MEDIUM FOR VIDEO PROCESSING

EP4736452A1EP 4736452 A1EP4736452 A1EP 4736452A1EP-4736452-A1

Abstract

Embodiments of the present disclosure provide a solution for video processing. A method for video processing is proposed. The method comprises: deriving, for a conversion between a video unit of a video and a bitstream of the video, a refined prediction sample of a prediction sample in the video unit by applying a function used in local illumination compensation (LIC) to the prediction sample, in response to that the video unit is coded with a target coding mode, wherein one or more parameters of the function are modified and the target coding mode is applied to at least one of: a prediction mode or a coding tool; and performing the conversion based on the refined prediction sample.

Inventors

  • WANG, YANG
  • ZHANG, KAI
  • HE, YUWEN
  • LIU, HONGBIN
  • ZHANG, LI

Assignees

  • DOUYIN VISION CO., LTD.
  • ByteDance Inc.

Dates

Publication Date
20260506
Application Date
20240627

Claims (20)

  1. A method for video processing, comprising: deriving, for a conversion between a video unit of a video and a bitstream of the video, a refined prediction sample of a prediction sample in the video unit by applying a function used in local illumination compensation (LIC) to the prediction sample, in response to that the video unit is coded with a target coding mode, wherein one or more parameters of the function are modified and the target coding mode is applied to at least one of: a prediction mode or a coding tool; and performing the conversion based on the refined prediction sample.
  2. The method of claim 1, wherein the prediction mode comprises at least one of: an intra prediction mode, an inter prediction mode, an intra block copy (IBC) prediction mode, or other prediction mode.
  3. The method of claim 1, wherein if the target coding mode is used, one or more coding tools are not used.
  4. The method of claim 3, wherein overlapped block motion compensation (OBMC) is not used.
  5. The method of claim 4, wherein an indication of OBMC is not signaled, if the target coding mode is used.
  6. The method of claim 1, wherein the target coding mode is not allowed to be used for bi-prediction.
  7. The method of claim 1, wherein at least one of: pre-defined adjustment parameter, signaled adjustment parameter or derived adjustment parameter is used for a block based reference picture derivation.
  8. The method of claim 8, wherein the pre-defined adjustment parameter is used for the block based reference picture derivation.
  9. The method of claim 8, wherein the pre-defined adjustment parameter is equal to one of: 1, -1, 2, -2, 3, or -3.
  10. The method of claim 1, wherein the target coding mode is not used during a block based reference picture derivation.
  11. The method of claim 1, wherein the target coding mode is applied to one or more inter coding tools, or wherein the target coding mode is not applied to one or more inter coding tools.
  12. The method of claim 11, wherein the one or more inter coding tools comprise at least one of: a combined inter and intra prediction (CIIP) , a variant of CIIP, a bi-prediction with coding unit-level weight (BCW) , a variant of BCW a merge mode with motion vector difference (MMVD) , a variant of MMVD, a template matching (TM) , a variant of TM, an affine, a variant of affine, a decoder side motion vector refinement (DMVR) , a variant of DMVR, a multi-pass DMVR, a prediction refinement with optical flow (PROF) , a variant of PROF, a bi-directional optical flow (BDOF) , a variant of BDOF, a sample based BDOF, an adaptive decoder-side motion vector refinement (ADMVR) , a variant of ADMVR, an OBMC, a variant of OBMC, a TM based OBMC, a multi-hypothesis prediction (MHP) , a variant of MHP, a geometric partitioning mode (GPM) , a variant of GPM, a bilateral matching advanced motion vector prediction (AMVP) -merge mode, a template matching AMVP-merge mode, an Inter convolutional cross-component model (CCCM) , an Inter cross-component prediction (CCP) merge mode, a reference picture resampling (RPR) , or a variant of RPR.
  13. The method of claim 12, wherein the CIIP comprises at least one of: CIIP-Planar, CIIP-template-based intra mode derivation (TIMD) , or CIIP-TM, or wherein the BCW comprises BCW index derived by TM, or wherein the MMVD comprises a TM based reordering for MMVD, or wherein the affine comprises at least one of: affine-MMVD, TM based reordering for affine MMVD, or wherein the GPM comprises at least one of: GPM-TM, GPM-MMVD, GPM-intra, GPM-Affine, GPM-IBC, regression-based GPM blending.
  14. The method of claim 12, wherein the target coding mode is not used for RPR.
  15. The method of claim 12, wherein if at least one of: BCW, adaptive motion vector resolution (AMVR) , or symmetric motion vector differences (SMVD) is used, a syntax element indicating whether the target coding mode is used is not signaled.
  16. The method of claim 12, wherein the target coding mode is not used for a cross-component coding tool which uses cross-component information to get the prediction or reconstruction.
  17. The method of claim 16, wherein the cross-component coding tool comprises at least one of: a cross-component linear model (CCLM) , a variant of CCLM, a multi-model CCLM (MMLM) , a variant of MMLM, a gradient linear model (GLM) , a variant of GLM, an inter convolutional cross-component model (CCCM) , a variant of inter CCCM, a CCP merge, or a variant of CCP merge.
  18. The method of claim 16, wherein the target coding mode is not used with the cross-component coding tool.
  19. The method of claim 18, wherein if the target coding mode is used for the cross-component coding tool, parameters of LIC for luma and/or chroma components are not adjusted.
  20. The method of claim 18, wherein if the target coding mode is used for the cross-component coding tool, LIC is disabled for luma and/or chroma components.

Description

METHOD, APPARATUS, AND MEDIUM FOR VIDEO PROCESSING FIELDS Embodiments of the present disclosure relates generally to video processing techniques, and more particularly, to local illumination compensation with slope adjustment. BACKGROUND In nowadays, digital video capabilities are being applied in various aspects of peoples' lives. Multiple types of video compression technologies, such as MPEG-2, MPEG-4, ITU-TH. 263, ITU-TH. 264/MPEG-4 Part 10 Advanced Video Coding (AVC) , ITU-TH. 265 high efficiency video coding (HEVC) standard, versatile video coding (VVC) standard, have been proposed for video encoding/decoding. However, coding efficiency of video coding techniques is generally expected to be further improved. SUMMARY Embodiments of the present disclosure provide a solution for video processing. In a first aspect, a method for video processing is proposed. The method comprises: deriving, for a conversion between a video unit of a video and a bitstream of the video, a refined prediction sample of a prediction sample in the video unit by applying a function used in local illumination compensation (LIC) to the prediction sample, in response to that the video unit is coded with a target coding mode, wherein one or more parameters of the function are modified and the target coding mode is applied to at least one of: a prediction mode or a coding tool; and performing the conversion based on the refined prediction sample. In this way, coding performance can be improved by adjusting the parameters. In a second aspect, an apparatus for video processing is proposed. The apparatus comprises a processor and a non-transitory memory with instructions thereon. The instructions upon execution by the processor, cause the processor to perform a method in accordance with the first aspect of the present disclosure. In a third aspect, a non-transitory computer-readable storage medium is proposed. The non-transitory computer-readable storage medium stores instructions that  cause a processor to perform a method in accordance with the first aspect of the present disclosure. In a fourth aspect, another non-transitory computer-readable recording medium is proposed. The non-transitory computer-readable recording medium stores a bitstream of a video which is generated by a method performed by an apparatus for video processing. The method comprises: deriving a refined prediction sample of a prediction sample in a video unit of the video by applying a function used in local illumination compensation (LIC) to the prediction sample, in response to that the video unit is coded with a target coding mode, wherein one or more parameters of the function are modified and the target coding mode is applied to at least one of: a prediction mode or a coding tool; and generating the bitstream of the video unit based on the refined prediction sample. In a fifth aspect, a method for storing a bitstream of a video is proposed. The method comprises: deriving a refined prediction sample of a prediction sample in a video unit of the video by applying a function used in local illumination compensation (LIC) to the prediction sample, in response to that the video unit is coded with a target coding mode, wherein one or more parameters of the function are modified and the target coding mode is applied to at least one of: a prediction mode or a coding tool; generating the bitstream of the video unit based on the refined prediction sample; and storing the bitstream in a non-transitory computer-readable recording medium. This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. BRIEF DESCRIPTION OF THE DRAWINGS Through the following detailed description with reference to the accompanying drawings, the above and other objectives, features, and advantages of example embodiments of the present disclosure will become more apparent. In the example embodiments of the present disclosure, the same reference numerals usually refer to the same components. Fig. 1 illustrates a block diagram that illustrates an example video coding system, in accordance with some embodiments of the present disclosure; Fig. 2 illustrates a block diagram that illustrates a first example video encoder, in accordance with some embodiments of the present disclosure; Fig. 3 illustrates a block diagram that illustrates an example video decoder, in accordance with some embodiments of the present disclosure; Fig. 4 illustrates an example of encoder block diagram; Fig. 5 illustrates 67 intra prediction modes; Fig. 6A and Fig. 6B illustrate reference samples for wide-angular intra prediction, respectively; Fig. 7 illustrates problem of discontinuity in case of directions beyond 45°; Fig. 8A and Fig. 8B illustrate MMVD S