Search

EP-4740465-A1 - EXTRAPOLATION INTRA PREDICTION MODEL INHERITANCE

EP4740465A1EP 4740465 A1EP4740465 A1EP 4740465A1EP-4740465-A1

Abstract

A method for using Extrapolation Intra Prediction (EIP) to code pixel blocks is provided. A video coder receives data to be encoded or decoded as a current block of pixels of a current picture of a video. The video coder inherits a previously used linear model for extrapolation intra prediction (EIP) that is derived based on samples in a template region neighboring a previously coded block, the template region occupying an area that extends beyond at least one of top and left edges of the previously coded block. The video coder applies the inherited linear model to generate a predictor of the current block by using reconstructed samples or predicted samples of the current block as input to a filter to generate samples of the predictor. The video coder encodes or decodes the current block by using the predictor.

Inventors

  • CHUANG, CHENG-YEN
  • CHUANG, TZU-DER
  • CHEN, YI-WEN
  • CHEN, CHING-YEH
  • HSU, CHIH-WEI
  • HUANG, YU-WEN

Assignees

  • MediaTek Inc.

Dates

Publication Date
20260513
Application Date
20240705

Claims (16)

  1. A video coding method comprising: receiving data to be encoded or decoded as a current block of pixels of a current picture of a video; inheriting a previously used linear model for extrapolation intra prediction (EIP) that is derived based on samples in a template region neighboring a previously coded block, the template region occupying an area that extends beyond at least one of top and left edges of the previously coded block; applying the inherited linear model to generate a predictor of the current block by using reconstructed samples or predicted samples of the current block as input to a filter to generate samples of the predictor; and encoding or decoding the current block by using the predictor.
  2. The video coder of claim 1, wherein inheriting the linear model comprises inheriting a set of filter coefficients.
  3. The video coder of claim 2, wherein inheriting the linear model comprises inheriting offset values to be applied to filter taps corresponding to the set of filter coefficients.
  4. The video coder of claim 2, wherein inheriting the linear model comprises inheriting maximum and minimum values for constraining the output of the linear model.
  5. The video coder of claim 1, wherein inheriting the linear model comprises inheriting information regarding a derivation of the linear model.
  6. The video coder of claim 5, wherein the information comprises a selection of a type of template region from multiple types of template region comprising above template region type, left template region type, and above-left template region type.
  7. The video coder of claim 5, wherein the information comprises a selection of a filter kernel type from multiple different filter kernel types having different shapes.
  8. The video coder of claim 1, wherein the linear model is inherited from a spatial neighbor of the current block.
  9. The video coder of claim 1, wherein the linear model is inherited from a non-adjacent spatial neighbor of the current block.
  10. The video coder of claim 1, wherein the inherited linear model is identified by using a history table that stores a history of one or more previously used linear models.
  11. The video coder of claim 1, wherein the linear model is inherited from a neighboring position of a corresponding position of the current block in a reference picture.
  12. The video coder of claim 11, wherein the neighboring position is constrained to be within a region in the reference picture that is defined based on the corresponding position of the current block in the reference picture.
  13. The video coder of claim 1, further comprising constructing a candidate list that includes one or more linear models, wherein inheriting the linear model comprises signaling or receiving a selection of the linear model from the constructed candidate list.
  14. An electronic apparatus comprising: a video coder circuit configured to perform operations comprising: receiving data to be encoded or decoded as a current block of pixels of a current picture of a video; inheriting a previously used linear model for extrapolation intra prediction (EIP) that is derived based on samples in a template region neighboring a previously coded block, the template region occupying an area that extends beyond at least one of top and left edges of the previously coded block; applying the inherited linear model to generate a predictor of the current block by using reconstructed samples or predicted samples of the current block as input to a filter to generate samples of the predictor; and encoding or decoding the current block by using the predictor.
  15. A video decoding method comprising: receiving data to be decoded as a current block of pixels of a current picture of a video; inheriting a previously used linear model for extrapolation intra prediction (EIP) that is derived based on samples in a template region neighboring a previously coded block, the template region occupying an area that extends beyond at least one of top and left edges of the previously coded block; applying the inherited linear model to generate a predictor of the current block by using reconstructed samples or predicted samples of the current block as input to a filter to generate samples of the predictor; and reconstructing the current block by using the predictor.
  16. A video encoding method comprising: receiving data to be encoded as a current block of pixels of a current picture of a video; inheriting a previously used linear model for extrapolation intra prediction (EIP) that is derived based on samples in a template region neighboring a previously coded block, the template region occupying an area that extends beyond at least one of top and left edges of the previously coded block; applying the inherited linear model to generate a predictor of the current block by using reconstructed samples or predicted samples of the current block as input to a filter to generate samples of the predictor; and encoding the current block by using the predictor.

Description

EXTRAPOLATION INTRA PREDICTION MODEL INHERITANCE CROSS REFERENCE TO RELATED PATENT APPLICATION (S) The present disclosure is part of a non-provisional application that claims the priority benefit of U.S. Provisional Patent Application No. 63/525, 432 filed on 7 July 2023. Content of above-listed applications are herein incorporated by reference. TECHNICAL FIELD The present disclosure relates generally to video coding. In particular, the present disclosure relates to methods of coding pixel blocks by extrapolation intra prediction (EIP) . BACKGROUND Unless otherwise indicated herein, approaches described in this section are not prior art to the claims listed below and are not admitted as prior art by inclusion in this section. High-Efficiency Video Coding (HEVC) is an international video coding standard developed by the Joint Collaborative Team on Video Coding (JCT-VC) . HEVC is based on the hybrid block-based motion-compensated DCT-like transform coding architecture. The basic unit for compression, termed coding unit (CU) , is a 2Nx2N square block of pixels, and each CU can be recursively split into four smaller CUs until the predefined minimum size is reached. Each CU contains one or multiple prediction units (PUs) . Versatile video coding (VVC) is the latest international video coding standard developed by the Joint Video Expert Team (JVET) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11. The input video signal is predicted from the reconstructed signal, which is derived from the coded picture regions. The prediction residual signal is processed by a block transform. The transform coefficients are quantized and entropy coded together with other side information in the bitstream. The reconstructed signal is generated from the prediction signal and the reconstructed residual signal after inverse transform on the de-quantized transform coefficients. The reconstructed signal is further processed by in-loop filtering for removing coding artifacts. The decoded pictures are stored in the frame buffer for predicting the future pictures in the input video signal. In VVC, a coded picture is partitioned into non-overlapped square block regions represented by the associated coding tree units (CTUs) . The leaf nodes of a coding tree correspond to the coding units (CUs) . A coded picture can be represented by a collection of slices, each comprising an integer number of CTUs. The individual CTUs in a slice are processed in raster-scan order. A bi-predictive (B) slice may be decoded using intra prediction or inter prediction with at most two motion vectors and reference indices to predict the sample values of each block. A predictive (P) slice is decoded using intra prediction or inter prediction with at most one motion vector and reference  index to predict the sample values of each block. An intra (I) slice is decoded using intra prediction only. A CTU can be partitioned into one or multiple non-overlapped coding units (CUs) using the quadtree (QT) with nested multi-type-tree (MTT) structure to adapt to various local motion and texture characteristics. A CU can be further split into smaller CUs using one of the five split types: quad-tree partitioning, vertical binary tree partitioning, horizontal binary tree partitioning, vertical center-side triple-tree partitioning, horizontal center-side triple-tree partitioning. Each CU contains one or more prediction units (PUs) . The prediction unit, together with the associated CU syntax, works as a basic unit for signaling the predictor information. The specified prediction process is employed to predict the values of the associated pixel samples inside the PU. Each CU may contain one or more transform units (TUs) for representing the prediction residual blocks. A transform unit (TU) is comprised of a transform block (TB) of luma samples and two corresponding transform blocks of chroma samples and each TB correspond to one residual block of samples from one color component. An integer transform is applied to a transform block. The level values of quantized coefficients together with other side information are entropy coded in the bitstream. The terms coding tree block (CTB) , coding block (CB) , prediction block (PB) , and transform block (TB) are defined to specify the 2-D sample array of one-color component associated with CTU, CU, PU, and TU, respectively. Thus, a CTU consists of one luma CTB, two chroma CTBs, and associated syntax elements. A similar relationship is valid for CU, PU, and TU. For each inter-predicted CU, motion parameters consisting of motion vectors, reference picture indices and reference picture list usage index, and additional information are used for inter-predicted sample generation. The motion parameter can be signalled in an explicit or implicit manner. When a CU is coded with skip mode, the CU is associated with one PU and has no significant residual coefficients, no coded motion vector delta or reference picture index. A merge mode is specif