Search

US-12627830-B2 - Image processing apparatus and method

US12627830B2US 12627830 B2US12627830 B2US 12627830B2US-12627830-B2

Abstract

The present disclosure relates to an image processing apparatus and an image processing method capable of improving processing efficiency with pipeline processing in encoding or decoding of a motion vector. In a motion vector encoding unit, such configuration is adopted that when a spatial prediction motion vector is derived according to AMVP or Merge mode, the use of a motion vector of a PU adjacent to a top right of a PU in question is prohibited. Therefore, the motion vector encoding unit performs encoding processing of a motion vector by using only motion vector information of B 1 , B 2 which are PUs located at Top with respect to the PU in question and A 0 , A 1 which are PUs located at Left with respect to the PU in question. The present disclosure can be applied to, for example, an image processing apparatus.

Inventors

  • Kazushi Sato

Assignees

  • SONY CORPORATION

Dates

Publication Date
20260512
Application Date
20230406
Priority Date
20111110

Claims (9)

  1. 1 . An image processing apparatus comprising: a circuitry configured to: set identification information for identifying whether or not to use merge mode for candidates of prediction motion vector; prohibit use of a motion vector of a top right block located adjacent to top right of a current block as one of the candidates of prediction motion vector of the merge mode based on the set identification information; generate a spatial prediction vector of the current block, using as one of the candidates of the prediction motion vector of the merge mode a motion vector adjacent in terms of space to the current block other than the motion vector of the top right block which is prohibited from being used, and set an index of the spatial prediction vector; and encode the motion vector of the current block, using the spatial prediction vector as a prediction motion vector of the current block, wherein the identification information is set in a unit in which control is to be performed, wherein the circuitry is further configured to: prohibit the use of the motion vector of the top right block in a maximum encoding unit; determine whether a border of the current block is a border of the maximum encoding unit; and prohibit the use of the motion vector of the top right block when the circuitry determines that the border of the current block is the border of the maximum encoding unit.
  2. 2 . The image processing apparatus according to claim 1 , wherein the circuitry is further configured to transmit a coded stream including the encoded motion vector.
  3. 3 . The image processing apparatus according to claim 2 , wherein the circuitry is further configured to transmit prediction mode information that is included in the coded stream.
  4. 4 . The image processing apparatus according to claim 1 , wherein the circuitry is further configured to perform generation processing of the spatial prediction vector with respect to the current block and generation processing of a spatial prediction vector with respect to a block subsequent to the current block in scan order.
  5. 5 . The image processing apparatus according to claim 4 , wherein the circuitry is further configured to generate the spatial prediction vector of the current block, using a motion vector of a first block which is a spatial adjacent block of the current block and which is located at a right end with a top block in surface contact with a top of the current block being as a target, and a motion vector of a second block other than the first block.
  6. 6 . The image processing apparatus according to claim 1 , wherein the identification information is set in a picture parameter set based on the control to be performed in a picture unit at an encoding side, wherein the identification information is set in a slice header based on the control to be performed in a slice unit, and wherein the identification information is set in a sequence parameter set based on the control to be performed in a sequence unit.
  7. 7 . An image processing method comprising: setting identification information for identifying whether or not to use merge mode for candidates of prediction motion vector; prohibiting use of a motion vector of a top right block located adjacent to top right of a current block as one of the candidates of prediction motion vector of the merge mode based on the set identification information; generating a spatial prediction vector of the current block, using as one of the candidates of the prediction motion vector of the merge mode a motion vector adjacent in terms of space to the current block other than the motion vector of the top right block which is prohibited from being used, and setting an index of the spatial prediction vector; encoding the motion vector of the current block, using the spatial prediction vector as a prediction motion vector of the current block, wherein the identification information is set in a unit in which control is to be performed; prohibiting the use of the motion vector of the top right block in a maximum encoding unit; determining whether a border of the current block is a border of the maximum encoding unit; and prohibiting the use of the motion vector of the top right block when it is determined that the border of the current block is the border of the maximum encoding unit.
  8. 8 . An image processing apparatus comprising: a circuitry configured to: prohibit use of a motion vector of a top right block located adjacent to top right of a current block as one of candidates of prediction motion vector of merge mode based on identification information for identifying whether or not to use the merge mode for the candidates of prediction motion vector; generate a spatial prediction vector of the current block, using as one of the candidates of the prediction motion vector of the merge mode a motion vector adjacent in terms of space to the current block other than the motion vector of the top right block which is prohibited from being used and an index of the spatial prediction vector; decode a motion vector of the current block, using the spatial prediction vector as a prediction motion vector of the current block; prohibit the use of the motion vector of the top right block in a maximum encoding unit; determine whether a border of the current block is a border of the maximum encoding unit; and prohibit the use of the motion vector of the top right block when the circuitry determines that the border of the current block is the border of the maximum encoding unit.
  9. 9 . An image encoding method comprising: prohibiting use of a motion vector of a top right block located adjacent to top right of a current block as one of candidates of prediction motion vector of merge mode based on identification information for identifying whether or not to use the merge mode for the candidates of prediction motion vector; generating a spatial prediction vector of the current block, using as one of the candidates of the prediction motion vector of the merge mode a motion vector adjacent in terms of space to the current block other than the motion vector of the top right block which is prohibited from being used and an index of the spatial prediction vector; decoding a motion vector of the current block, using the spatial prediction vector as a prediction motion vector of the current block; prohibiting the use of the motion vector of the top right block in a maximum encoding unit; determining whether a border of the current block is a border of the maximum encoding unit; and prohibiting the use of the motion vector of the top right block when it is determined that the border of the current block is the border of the maximum encoding unit.

Description

CROSS REFERENCE TO PRIOR APPLICATION This application is a continuation of U.S. patent application Ser. No. 16/386,739 (filed on Apr. 17, 2019), which is a continuation of U.S. patent application Ser. No. 14/348,097 (filed on Mar. 28, 2014 and issued as U.S. Pat. No. 10,616,599 on Apr. 7, 2020), which is a National Stage Patent Application of PCT International Patent Application No. PCT/JP2012/078427 (filed on Nov. 2, 2012) under 35 U.S.C. § 371, which claims priority to Japanese Patent Application Nos. 2011-247489 (filed on Nov. 11, 2011) and 2011-246543 (filed on Nov. 10, 2011), which are all hereby incorporated by reference in their entirety. TECHNICAL FIELD The present disclosure relates to an image processing apparatus and an image processing method, and more particularly, relates to an image processing apparatus and an image processing method capable of improving the processing efficiency by pipeline processing in encoding or decoding of a motion vector. BACKGROUND ART In recent years, image information is treated as digital information, and at that occasion, an apparatus is becoming widely prevalent that compresses and encodes an image by employing a coding method for performing compression by orthogonal transformation and motion compensation such as discrete cosine transform using redundancy unique to image information for the purpose of highly efficient transmission and accumulation of information. Examples of the coding methods include MPEG (Moving Picture Experts Group) and the like. In particular, MPEG2 (ISO/IEC 13818-2) is defined as a general-purpose image coding method, and is a standard that covers both of interlaced scanning images and progressive scanning images, and standard resolution images and high resolution images. For example, MPEG2 is now widely used for wide range of applications such as professional use and consumer use. For example, in a case of an interlaced scanning image of a standard resolution of 720 by 480 pixels, the amount of codes (bit rate) of 4 to 8 Mbps is allocated by using the MPEG2 compression method. Further, for example, in a case of an interlaced scanning image of a high resolution of 1920 by 1088 pixels, the amount of codes (bit rate) of 18 to 22 Mbps is allocated by using the MPEG2 compression method. Accordingly, high compression rate and high image quality can be achieved. MPEG2 is mainly targeted for high image quality coding suitable for broadcasting, but does not support coding method of a less amount of codes (bit rate) than MPEG1. In other words, MPEG2 does not support higher compression rate. As portable terminals become widely prevalent, needs for such coding methods are considered to grow in the future, and in order to respond to such needs, MPEG 4 coding method has been standardized. With regard to image coding method, the specification is admitted as ISO/IEC14496-2 in international standard on December, 1998. In the schedule of standardization, on March, 2003, H.264 and MPEG-4 Part 10 (Advanced Video Coding, hereinafter referred to as H.264/AVC) was made into an international standard. Further, as an expansion of the H.264/AVC, FRExt (Fidelity Range Extension) including an encoding tool required for professional use such as RGB, 4:2:2, 4:4:4 and even 8×8 DCT and quantization matrix specified in the MPEG-2 was standardized on February, 2005. Accordingly, there was made a coding method capable of expressing even film noise included in a movie in a preferable manner using the H.264/AVC, and it is now being used for a wide range of applications such as Blu-Ray Disc (trademark). However, recently, needs for still higher compression rate encoding has been enhanced, e.g., compressing an image of about 4000 by 2000 pixels which is four times the high-definition image, and distributing a high-definition image in an environment of a limited transmission capacity such as the Internet. For this reason, in VCEG (=Video Coding Expert Group) under ITU-T explained above, discussions about improvement of the encoding efficiency have been continuously conducted. As one of such encoding efficiency improvements, in order to improve encoding of the motion vector using median prediction according to the AVC, adaptive use of any of not only “Spatial Predictor” derived from the median prediction defined in the AVC, but also “Temporal Predictor” and “Spatio-Temporal Predictor” as prediction motion vector information (hereinafter also referred to as MV Competition (MVCompetition)) has been suggested (for example, see Non-Patent Document 1). It should be noted that, in the AVC, when prediction motion vector information is selected, a cost function value in High Complexity Mode or Low Complexity Mode implemented in the reference software of the AVC which is called JM (Joint Model) is used. More specifically, a cost function value in a case where the prediction motion vector information is used is calculated, and the optimum prediction motion vector information is selected. I