CN-122029811-A - Method, apparatus and medium for video processing
Abstract
Embodiments of the present disclosure provide a solution for video processing. A method for video processing is presented. The method includes obtaining a chroma fusion candidate for a current block of a video for a conversion between the current block and a bitstream of the video, and performing the conversion based on the chroma fusion candidate.
Inventors
- DENG ZHIPIN
- ZHANG KAI
- ZHANG LI
Assignees
- 抖音视界有限公司
- 字节跳动有限公司
Dates
- Publication Date
- 20260512
- Application Date
- 20241010
- Priority Date
- 20231011
Claims (20)
- 1. A method for video processing, comprising: acquiring chroma fusion candidates for a current block of video for conversion between the current block and a bitstream of the video, and The conversion is performed based on the chroma fusion candidates.
- 2. The method of claim 1, wherein the list of chroma fusion candidates is generated based on decoding information.
- 3. The method of claim 2, wherein chroma fusion candidates in the list of chroma fusion candidates are generated based on at least one of the following patterns: an intra convolution cross-component model (CCCM) mode, An inter-frame CCCM mode is used, CCCM modes with Multiple Downsampling Filters (MDFs), Gradient linear-convolution cross-component model (GL-CCCM) mode, A cross-component linear model (CCLM) mode, A Gradient Linear Model (GLM) mode, A local lifting cross-component prediction (LBCCP) mode, A single model of the device is provided, A plurality of models of the model are arranged, The decoder-side intra mode derives DIMD modes, Template-based intra mode derivation (TIMD) of modes, The Derivation Mode (DM), Based on the pattern of the linear model, Modes based on non-linear models, or Based on the pattern of the convolution model.
- 4. The method of claim 2, wherein what type of chroma fusion candidate is included in the list of chroma fusion candidates is determined based on a predetermined rule.
- 5. The method of claim 1, wherein the chroma fusion candidates are determined in real-time based on at least one of a reference region or a training region.
- 6. The method of claim 5, wherein at least one of the chroma fusion filter or the chroma fusion model is derived based on a process that minimizes a difference between a value of a luma sample and a value of a chroma sample of a reference, a training area, or a block.
- 7. The method of claim 6, wherein the chroma fusion model is applied to luma reconstruction samples of the reference, the training area, or the block to obtain resulting model estimation samples.
- 8. The method of claim 7, wherein the minimizing process estimates values of samples and chroma reconstruction samples based on the resulting model, the reference, the training area, or the block.
- 9. The method of claim 6, wherein one of the reference, the training area, or the block, or Wherein one of the reference, the training area, or the block is not adjacent to the current block, or Wherein one of the reference, the training area, or the block is temporally co-located with the current block.
- 10. The method of claim 6, wherein one of the reference, the training area, or the block is derived based on a block vector.
- 11. The method of claim 6, wherein one of the reference, the training area, or the block is derived based on a motion vector.
- 12. The method of claim 5, wherein at least one of a chroma fusion filter, or a chroma fusion model, is derived based on a process that minimizes a difference between a reference template and a current template.
- 13. The method of claim 12, wherein the reference template is adjacent to a reference block, or Wherein the reference template is not adjacent to a reference block.
- 14. The method of claim 12, wherein the current template is adjacent to the current block, or Wherein the current template is not adjacent to the current block.
- 15. The method of claim 12, wherein the model is applied to reference template samples to obtain resulting model estimation samples.
- 16. The method of claim 15, wherein the minimizing process estimates values of samples and reconstructed current template samples based on the resulting model.
- 17. The method of claim 5, wherein at least one of a chroma fusion filter, or a chroma fusion model, is calculated based on the training set.
- 18. The method of claim 17, wherein the training set is determined based on a set of samples from at least one of the current block, a Block Vector (BV) guided reference block, a Motion Vector (MV) guided reference block, a template, a non-neighboring block, a temporal homoblock, a temporal block neighboring a homoblock, or a temporal block non-neighboring the homoblock.
- 19. The method of claim 18, wherein multiple types of samples are used to determine the training set.
- 20. The method of claim 18, wherein training samples for computing coefficients of a model are derived from a block that was encoded and decoded prior to the current block, wherein the block was encoded and decoded by the model.
Description
Method, apparatus and medium for video processing Technical Field Embodiments of the present disclosure relate generally to video processing techniques and, more particularly, to cross-component prediction in coding. Background Today, digital video capabilities are being applied to various aspects of a person's life. Various types of video compression techniques have been proposed for video encoding/decoding, such as the MPEG-2, MPEG-4, ITU-T H.263, ITU-T H.264/MPEG-4 part 10 Advanced Video Codec (AVC), ITU-T H.265 High Efficiency Video Codec (HEVC) standard, the multifunctional video codec (VVC) standard. However, the codec efficiency of video codec technology is generally expected to be further improved. Disclosure of Invention Embodiments of the present disclosure provide a solution for video processing. In a first aspect, a method for video processing is presented. The method includes obtaining a chroma fusion candidate for a current block of a video for a conversion between the current block and a bitstream of the video, and performing the conversion based on the chroma fusion candidate. The method according to the first aspect of the present disclosure advantageously improves coding efficiency and performance compared to conventional solutions by obtaining chroma fusion candidates for the current block. In a second aspect, another method for video processing is presented. The method includes applying LBCCP filters to CCP modes or CCP candidates for conversion between a current block of video and a bitstream of video based on at least one of template cost or a block preceding the current block, and performing conversion based on LBCCP filters. The method according to the second aspect of the present disclosure advantageously improves codec efficiency and performance compared to conventional solutions by applying LBCCP filters to CCP modes or CCP candidates. In a third aspect, an apparatus for video processing is presented. The apparatus includes a processor and a non-transitory memory having instructions thereon. The instructions, when executed by a processor, cause the processor to perform a method according to the first or second aspect of the present disclosure. In a fourth aspect, a non-transitory computer readable storage medium is presented. The non-transitory computer readable storage medium stores instructions that cause a processor to perform a method according to the first or second aspect of the present disclosure. In a fifth aspect, another non-transitory computer readable recording medium is presented. The non-transitory computer readable recording medium stores a bitstream of video generated by a method performed by a video processing apparatus. The method includes obtaining a chroma fusion candidate for a current block of a video and generating a bitstream based on the chroma fusion candidate. In a sixth aspect, another non-transitory computer readable recording medium is presented. The non-transitory computer readable recording medium stores a method of video performed by a video processing apparatus to generate a bitstream. The method includes applying LBCCP filters to CCP modes or CCP candidates based on at least one of a template cost or a block preceding a current block of video, and generating a bitstream based on LBCCP filters. In a seventh aspect, a method for storing a bitstream of video is presented. The method includes obtaining a chroma fusion candidate for a current block of the video, generating a bitstream based on the chroma fusion candidate, and storing the bitstream in a non-transitory computer-readable recording medium. In an eighth aspect, a method for storing a bitstream of video is presented. The method includes applying LBCCP a filter to a CCP mode or CCP candidate based on at least one of a template cost or a block preceding a current block of video, generating a bitstream based on LBCCP the filter, and storing the bitstream in a non-transitory computer-readable recording medium. This summary is intended to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Drawings The above and other objects, features and advantages of the exemplary embodiments of the present disclosure will become more apparent by the following detailed description with reference to the accompanying drawings. In example embodiments of the present disclosure, like reference numerals generally refer to like components. FIG. 1 illustrates a block diagram of an example video codec system according to some embodiments of the present disclosure; fig. 2 illustrates a block diagram of a first example video encoder, according to some embodiments of the present disclosure; fig. 3 illustrates a block diagram of an example video decoder, according to some em