CN-122003869-A - Method and device for extrapolating intra-frame prediction model inheritance model in video coding and decoding
Abstract
A video encoding and decoding method and apparatus using an extrapolated intra prediction (Extrapolated Intra Prediction; EIP) mode correlation mode are disclosed. According to the method, input data associated with a current block is received, wherein the input data comprises pixel data to be encoded at an encoder side or data associated with the current block to be decoded at a decoder side. The current block is encoded or decoded using a non-EIP mode. One or more reference blocks pointed to by one or more motion vectors or one or more block vectors of the current block are determined. If the one or more reference blocks have target EIP information, the target EIP information is copied from the one or more reference blocks and stored into the current block, wherein the target EIP information stored in the current block is accessed by one or more subsequent blocks to derive prediction information.
Inventors
- ZENG XINYI
- ZHUANG ZHENGYAN
- CHEN QIWEN
- ZHUANG ZIDE
- CHEN QINGYE
- XU ZHIWEI
- HUANG YUWEN
Assignees
- 联发科技股份有限公司
Dates
- Publication Date
- 20260508
- Application Date
- 20241012
- Priority Date
- 20231012
Claims (20)
- 1. A video encoding and decoding method, comprising: Receiving input data associated with a current block, wherein the input data comprises pixel data to be encoded at an encoder side or data associated with the current block to be decoded at a decoder side; encoding or decoding the current block using a non-extrapolated intra prediction mode; Determining one or more reference blocks pointed to by one or more motion vectors of the current block or one or more block vectors of the current block, and If the one or more reference blocks have target extrapolated intra-prediction information, the target extrapolated intra-prediction information is copied and stored into the current block, where the target extrapolated intra-prediction information stored in the current block is accessed by one or more subsequent blocks to derive prediction information.
- 2. The method of claim 1, wherein the one or more block vectors are associated with a current luma block of the current block if the current block is copied in an intra block or an intra template matches a prediction mode codec.
- 3. The method of claim 1, wherein if the one or more reference blocks located according to the one or more block vectors or the one or more motion vectors are not extrapolated intra-prediction codecs, but first extrapolated intra-prediction information is stored in the one or more reference blocks, the stored first extrapolated intra-prediction information is used as the target extrapolated intra-prediction information.
- 4. The method of claim 1, wherein the first extrapolated intra-prediction information is used as the target extrapolated intra-prediction information if the one or more reference blocks correspond to a plurality of reference blocks and only one of the plurality of reference blocks has the first extrapolated intra-prediction information.
- 5. The method of claim 1, wherein if a plurality of reference blocks have first extrapolated intra-prediction information, the target extrapolated intra-prediction information is derived by combining extrapolated intra-prediction model parameters corresponding to all or part of the plurality of reference blocks.
- 6. The method of claim 1, wherein if a plurality of reference blocks have first extrapolated intra-prediction information, the target extrapolated intra-prediction information is copied from the first extrapolated intra-prediction information of target reference blocks of the plurality of reference blocks according to one or more predefined rules.
- 7. The method of claim 6, wherein a reference block extrapolated intra-prediction codec is selected as the target reference block of the plurality of reference blocks.
- 8. The method of claim 6, wherein an intra-frame codec, an inter-frame codec, or an intra-block copy codec reference block is selected as the target reference block for the plurality of reference blocks.
- 9. The method of claim 6, wherein a reference block having a shortest distance from the current block is selected as the target reference block of the plurality of reference blocks according to a distance metric.
- 10. A method as claimed in claim 9, wherein the distance measure corresponds to euclidean distance, manhattan distance, minkowski distance, horizontal distance or vertical distance.
- 11. The method of claim 6, wherein a reference block having a minimum distortion compared to the current block is selected as the target reference block of the plurality of reference blocks according to a distortion metric.
- 12. The method of claim 11, wherein the distortion metric corresponds to a distortion metric between predicted samples and reconstructed samples of a template of the current block.
- 13. The method of claim 1, wherein if the current block is inter-coded in bi-prediction and both reference blocks located by the one or more motion vectors of the current block store extrapolated intra-prediction information, then selecting one of the reference blocks according to a set of predefined rules.
- 14. The method of claim 13, wherein the one or more motion vectors of the current block are located in a center or upper left corner of the current block, a motion vector prediction of the current block, or a combination of motion vector prediction and motion vector difference of the current block.
- 15. The method of claim 13, wherein the set of predefined rules includes a dependency of picture order count distance between a reference picture associated with the reference block and a current picture.
- 16. The method of claim 13, wherein the set of predefined rules includes dependencies of quantization parameter values of the reference block.
- 17. The method of claim 13, wherein the location of the target block located by one motion vector is limited to a co-located codec tree cell line in the reference picture of the current codec tree cell line.
- 18. The method of claim 17, wherein if the location of the target block is above the location codec tree cell line, the location is mapped to a top line of the location codec tree cell line and extrapolated intra-prediction information of the mapped location is copied and stored into the current block.
- 19. The method of claim 17, wherein if the location of the target block is below the location codec tree cell line, the location is mapped to a bottom line of the location codec tree cell line, and extrapolated intra-prediction information of the mapped location is copied and stored into the current block.
- 20. A video codec device comprising one or more electronic circuits or processors configured to: Receiving input data associated with a current block, wherein the input data comprises pixel data to be encoded at an encoder side or data associated with the current block to be decoded at a decoder side; encoding or decoding the current block using a non-extrapolated intra prediction mode; Determining one or more reference blocks pointed to by one or more motion vectors of the current block or one or more block vectors of the current block, and If the one or more reference blocks have target extrapolated intra-prediction information, the target extrapolated intra-prediction information is copied and stored into the current block, where the target extrapolated intra-prediction information stored in the current block is accessed by one or more subsequent blocks to derive prediction information.
Description
Method and device for extrapolating intra-frame prediction model inheritance model in video coding and decoding Cross reference The present invention is a non-provisional application of U.S. provisional patent application No. 63/589,656 (filed on 12 th month 10 of 2023) and claims priority. The U.S. provisional patent application is incorporated herein by reference in its entirety. Technical Field The present invention relates to video coding and decoding systems using an extrapolated intra prediction (Extrapolation Intra Prediction; EIP) mode. In particular, the present invention relates to copying EIP information from blocks associated with EIP information and storing the EIP information in a current block when the current block is not encoded with EIP, wherein the EIP information is accessed by one or more subsequent blocks to derive prediction information. Background The multifunctional video codec (VERSATILE VIDEO CODING; VVC) is the latest international video codec standard developed by the international telecommunication union video codec expert group (ITU-T Video Coding Experts Group; ITU-T VCEG) and the international organization for standardization/international electrotechnical commission moving picture expert group (ISO/IEC Moving Picture Experts Group; ISO/IEC MPEG) in combination with the video expert group (Joint Video Experts Team; JVET). The standard has been published as an ISO standard, which is ISO/IEC 23090-3:2021 published in 2021, month 2, information technology-codec representation of immersive media-part 3-multifunctional video codec. VVC is developed on the basis of its predecessor (HIGH EFFICIENCY Video Coding (HEVC)), to improve Coding efficiency by adding more Coding tools, and to be able to handle various types of Video sources including three-dimensional (3D) Video signals. Fig. 1A illustrates an example adaptive inter/intra video coding system that includes loop processing. For intra prediction 110, the prediction data is derived based on previously encoded video data in the current picture. For inter prediction 112, motion estimation (Motion Estimation; ME) is performed at the encoder side and motion compensation (Motion Compensation; MC) is performed according to the results of ME to provide prediction data derived from other pictures and motion data. The switch 114 selects either intra prediction 110 or inter prediction 112, and the selected prediction data is provided to adder 116 to form a prediction error, also referred to as a residual. The prediction error is then processed by Transform (T) 118, followed by Quantization (Q) 120. The transformed and quantized residual is then encoded by entropy encoder (Entropy Encoder) 122 for inclusion in a video bitstream corresponding to the compressed video data. The bitstream associated with the transform coefficients is then packed with side information, such as motion and codec modes associated with intra and inter prediction, and other information, such as parameters associated with loop filters applied to the underlying image region, etc. As shown in fig. 1A, side information related to intra prediction 110, inter prediction 112, and loop filter 130 is provided to entropy encoder 122. When inter prediction modes are used, the reference picture or pictures must also be reconstructed at the encoder side. Thus, the transformed and quantized residual is processed through inverse quantization (Inverse Quantization; IQ) 124 and inverse transformation (Inverse Transformation; IT) 126 to recover the residual. The residual is then added back to the prediction data 136 at Reconstruction (REC) 128 to reconstruct the video data. The reconstructed video data may be stored in a reference picture buffer (REFERENCE PICTURE BUFFER) 134 and used for prediction of other frames. As shown in fig. 1A, input video data undergoes a series of processes in an encoding system. The reconstructed video data from the REC 128 may be subject to various impairments due to a series of processing. Therefore, loop filter 130 is typically applied to the reconstructed video data to improve video quality before it is stored in reference image buffer 134. For example, a deblocking filter (Deblocking Filter; DF), a sample adaptive Offset (SAMPLE ADAPTIVE Offset; SAO), and an adaptive loop filter (Adaptive Loop Filter; ALF) may be used. Loop filter information may need to be included in the bitstream in order for the decoder to be able to correctly recover the required information. Thus, loop filter information is also provided to the entropy encoder 122 to incorporate the bitstream. In fig. 1A, loop filter 130 is applied to the reconstructed video before the reconstructed samples are stored in reference image buffer 134. The system in fig. 1A is intended to show an example structure of a typical video encoder. It may correspond to HEVC system, VP8, VP9, h.264, or VVC. As shown in fig. 1B, the decoder may use the same or partially the same functional blocks as the en