Search

US-20260129240-A1 - METHOD, APPARATUS, AND MEDIUM FOR VIDEO PROCESSING

US20260129240A1US 20260129240 A1US20260129240 A1US 20260129240A1US-20260129240-A1

Abstract

Embodiments of the present disclosure provide a solution for video processing. A method for video processing is proposed. In the method, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) generated prediction is determined for the current video block. The CCP generated prediction is inherited from a CCP candidate. Filtering information regarding the CCP generated prediction is determined. The filtering information comprises at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction. The conversion is generated based on the filtering information.

Inventors

  • Kai Zhang
  • Li Zhang

Assignees

  • BYTEDANCE INC.

Dates

Publication Date
20260507
Application Date
20251231

Claims (20)

  1. 1 . A method for video processing, comprising: determining, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) generated prediction for the current video block, the CCP generated prediction being inherited from a CCP candidate; determining filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and performing the conversion based on the filtering information.
  2. 2 . The method of claim 1 , wherein the filtering process comprises a 3×3 filtering for a local-boosting CCP (LB-CCP); or wherein the filtering process comprises a 3×3 filtering for a local-boosting CCP (LB-CCP), and a filtering flag of the 3×3 filtering is inherited by the current video block.
  3. 3 . The method of claim 1 , wherein associated information of a CCP-coded block comprises the filtering information.
  4. 4 . The method of claim 3 , wherein the filtering information is stored in a unit block, or a history-based table.
  5. 5 . The method of claim 3 , further comprising: generating a further CCP candidate in a CCP candidate list of the current video block based on the filtering information.
  6. 6 . The method of claim 1 , further comprising: comparing first filtering information of a first CCP or potential candidate with second filtering information of a second CCP or potential candidate.
  7. 7 . The method of claim 6 , wherein the first filtering information is different from the second filtering information, and the first CCP or potential candidate is different from the second CCP or potential candidate, or wherein the first filtering information and the second filtering information are ignored.
  8. 8 . The method of claim 1 , wherein the filtering information is used for the conversion based on a type of the CCP candidate associated with the filtering information.
  9. 9 . The method of claim 8 , wherein the type of the CCP candidate is a multi-model CCP type, and the filtering information is used for the conversion.
  10. 10 . The method of claim 1 , wherein whether to and/or how to apply the filtering process on the CCP generated prediction is based on inherited filtering information.
  11. 11 . The method of claim 10 , wherein the inherited filtering information comprises an inherited filtering flag, the CCP generated prediction is filtered based on the inherited filtering flag being true, and the CCP generated prediction is not filtered based on the inherited filtering flag being false, and/or wherein the inherited filtering information is stored and used by a further block coded after the current video block.
  12. 12 . The method of claim 1 , wherein a syntax element in the bitstream is binarized as one of: a flag, a fixed length code, an exponential Golomb(x)(EG(x)) code, a unary code, a truncated unary code, or a truncated binary code, and/or wherein the syntax element is signed or unsigned.
  13. 13 . The method of claim 1 , wherein a syntax element in the bitstream is coded with at least one context model, or bypass coded, and/or wherein the syntax element is included in the bitstream based on a condition that a function associated with the syntax element is applicable, or wherein the syntax element is included in the bitstream if a dimension of the current video block satisfies a condition, the dimension of the current video block comprising at least one of: a width of the current video block, or a height of the current video block.
  14. 14 . The method of claim 12 , wherein the syntax element is at one of: a block level, a sequence level, a group of pictures level, a picture level, a slice level, or a tile group level, and/or wherein the syntax element is in one of the following coding structures: a coding tree unit (CTU), a coding unit (CU), a transform unit (TU), a prediction unit (PU), a coding tree block (CTB), a coding block (CB), a transform block (TB), a prediction block (PB), a sequence header, a picture header, a sequence parameter set (SPS), a Video Parameter Set (VPS), a decoded parameter set (DPS), Decoding Capability Information (DCI), a Picture Parameter Set (PPS), an Adaptation Parameter Set (APS), a slice header or a tile group header.
  15. 15 . The method of claim 1 , wherein information regarding whether to and/or how to apply the method is included in the bitstream, or wherein the method is used in a coding tool requiring a chroma fusion.
  16. 16 . The method of claim 15 , wherein the information is indicated at one of: a sequence level, a group of pictures level, a picture level, a slice level or a tile group level, and/or wherein the information is indicated in one of the following coding structures: a coding tree unit (CTU), a coding unit (CU), a transform unit (TU), a prediction unit (PU), a coding tree block (CTB), a coding block (CB), a transform block (TB), a prediction block (PB), a sequence header, a picture header, a sequence parameter set (SPS), a Video Parameter Set (VPS), a decoded parameter set (DPS), Decoding Capability Information (DCI), a Picture Parameter Set (PPS), an Adaptation Parameter Set (APS), a slice header or a tile group header, and/or wherein the information is based on coded information of the current video block, wherein the coded information comprises at least one of: a block size, a colour format, a single or dual tree partitioning, a colour component, a slice type, or a picture type.
  17. 17 . The method of claim 1 , wherein the conversion includes encoding the current video block into the bitstream, or wherein the conversion includes decoding the current video block from the bitstream.
  18. 18 . An apparatus for video processing comprising a processor and a non-transitory memory with instructions thereon, wherein the instructions upon execution by the processor, cause the processor to: determine, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) generated prediction for the current video block, the CCP generated prediction being inherited from a CCP candidate; determine filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and perform the conversion based on the filtering information.
  19. 19 . A non-transitory computer-readable storage medium storing instructions that cause a processor to: determine, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) generated prediction for the current video block, the CCP generated prediction being inherited from a CCP candidate; determine filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and perform the conversion based on the filtering information.
  20. 20 . A non-transitory computer-readable recording medium storing a bitstream of a video which is generated by a method performed by an apparatus for video processing, wherein the method comprises: determining a cross component prediction (CCP) generated prediction for a current video block of the video, the CCP generated prediction being inherited from a CCP candidate; determining filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and generating the bitstream based on the filtering information.

Description

CROSS REFERENCE This application is a continuation of International Application No. PCT/US2024/036764, filed on Jul. 3, 2024, which claims the benefit of U.S. Application No. 63/524,917, filed on Jul. 4, 2023. The entire contents of these applications are hereby incorporated by reference in their entireties. FIELD Embodiments of the present disclosure relates generally to video processing techniques, and more particularly, to cross component prediction (CCP) model. BACKGROUND In nowadays, digital video capabilities are being applied in various aspects of peoples' lives. Multiple types of video compression technologies, such as MPEG-2, MPEG-4, ITU-TH.263, ITU-TH.264/MPEG-4 Part 10 Advanced Video Coding (AVC), ITU-TH.265 high efficiency video coding (HEVC) standard, versatile video coding (VVC) standard, have been proposed for video encoding/decoding. However, coding efficiency of video coding techniques is generally expected to be further improved. SUMMARY Embodiments of the present disclosure provide a solution for video processing. In a first aspect, a method for video processing is proposed. The method comprises: determining, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) generated prediction for the current video block, the CCP generated prediction being inherited from a CCP candidate; determining filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and performing the conversion based on the filtering information. The method in accordance with the first aspect of the present disclosure determines whether to or how to apply filtering on the CCP generation inherited from the CCP candidate. The coding effectiveness and coding efficiency can thus be improved. In a second aspect, another method for video processing is proposed. The method comprises: determining, for a conversion between a current video block of a video and a bitstream of the video, a cross component prediction (CCP) candidate of the current video block; applying a decoder-derived determination to the CCP candidate; and performing the conversion based on the applying. The method in accordance with the second aspect of the present disclosure applies the decoder-derived determination to the CCP candidate. The coding effectiveness and coding efficiency can thus be improved. In a third aspect, an apparatus for video processing is proposed. The apparatus comprises a processor and a non-transitory memory with instructions thereon. The instructions upon execution by the processor, cause the processor to perform a method in accordance with the first, or second aspect of the present disclosure. In a fourth aspect, a non-transitory computer-readable storage medium is proposed. The non-transitory computer-readable storage medium stores instructions that cause a processor to perform a method in accordance with the first, or second aspect of the present disclosure. In a fifth aspect, another non-transitory computer-readable recording medium is proposed. The non-transitory computer-readable recording medium stores a bitstream of a video which is generated by a method performed by an apparatus for video processing. The method comprises: determining a cross component prediction (CCP) generated prediction for a current video block of the video, the CCP generated prediction being inherited from a CCP candidate; determining filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; and generating the bitstream based on the filtering information. In a sixth aspect, another non-transitory computer-readable recording medium is proposed. The non-transitory computer-readable recording medium stores a bitstream of a video which is generated by a method performed by an apparatus for video processing. The method comprises: determining a cross component prediction (CCP) generated prediction for a current video block of the video, the CCP generated prediction being inherited from a CCP candidate; determining filtering information regarding the CCP generated prediction, the filtering information comprising at least one of: whether to apply a filtering process on the CCP generated prediction, or how to apply the filtering process on the CCP generated prediction; generating the bitstream based on the filtering information; and storing the bitstream in a non-transitory computer-readable recording medium. In a seventh aspect, a method for storing a bitstream of a video is proposed. The method comprises: determining a cross component prediction (CCP) candidate of a current video block of the video; applying a