BR-112018009422-B1 - METHOD AND DEVICE FOR ADAPTING DECODED VIDEO CONTENT INTO A PROCESSOR-READABLE MEDIUM AND NON-TRANSIENTIAL STORAGE MEDIUM
Abstract
METHOD AND DEVICE FOR ADAPTING DECODED VIDEO CONTENT TO THE CHARACTERISTICS OF A DISPLAY FROM ELEMENTARY STREAMS. This disclosure relates to a method and device for adapting decoded video content from elementary streams to the characteristics of a display from at least one type of metadata that provides information regarding said elementary streams. This method comprises: - obtaining (102) additional information (HDR DESCR.) indicating the presence of a particular type of metadata; - determining whether said decoded video content from elementary streams can be displayed on said display (11) from said additional information (HDRDESCR.) and the display characteristics (EDID); and - if the said video content decoded from elementary streams is determined to be displayable, select (105) a process from the said additional information and the display characteristics and adapt (106) the video content according to the selected process.
Inventors
- Philippe Bordes
- Pierre Andrivon
- Edouard Francois
Assignees
- INTERDIGITAL VC HOLDINGS, INC
Dates
- Publication Date
- 20260317
- Application Date
- 20161104
- Priority Date
- 20151109
Claims (7)
- 1. A method for adapting video content decoded from elementary streams to the characteristics of a display based on at least one type of metadata that provides information about said elementary streams, CHARACTERIZED in that the method comprises: obtaining video data comprising a transport layer and at least one encapsulated elementary video stream; obtaining, from the transport layer, additional information indicating the presence of a particular type of metadata and descriptive information of the High Dynamic Range properties of at least one encapsulated elementary video stream, wherein the additional information comprises descriptive parameters of color primaries, transfer characteristics and matrix coefficients; obtaining display description data comprising display characteristics, wherein the characteristics comprise one or more formats supported by the display; determining whether said video content decoded from elementary streams is displayable on said display based on said additional information and the display characteristics; If the aforementioned video content decoded from elementary streams is determined to be displayable, select a process from the aforementioned additional information and the characteristics of the display and adapt the video content decoded from elementary streams according to the selected process, wherein the aforementioned additional information also indicates whether at least one particular type of metadata is present within the aforementioned signal for a particular duration of the aforementioned elementary streams.
- 2. Method, according to claim 1, CHARACTERIZED in that, when said at least one particular type of said metadata belongs to a set of dynamic metadata, said additional information comprises a first update information belonging to the group comprising: a maximum update rate of dynamic metadata; and a minimum time interval after which the dynamic metadata will be updated.
- 3. Method, according to claim 1, CHARACTERIZED in that, when said at least one particular type of said metadata belongs to a set of dynamic metadata, said additional information comprises a second update information belonging to the group comprising: a minimum update rate for dynamic metadata; and a maximum time interval after which the dynamic metadata will be updated.
- 4. Method, according to claim 1, CHARACTERIZED in that said encoded video content is High Dynamic Range video content, and said metadata is High Dynamic Range metadata.
- 5. Device for adapting video content decoded from elementary streams to the characteristics of a display based on at least one type of metadata that provides information regarding said elementary streams, the device CHARACTERIZED by the fact that it comprises a processor configured to: obtain video data comprising a transport layer and at least one encapsulated elementary video stream; obtain, from the transport layer, additional information indicating the presence of a particular type of metadata and descriptive information of the High Dynamic Range properties of at least one encapsulated elementary video stream, wherein the additional information comprises descriptive parameters of color primaries, transfer characteristics and matrix coefficients; obtain display description data comprising display characteristics, wherein the characteristics comprise one or more formats supported by the display; determine whether said video content decoded from elementary streams is displayable on said display based on said additional information and the display characteristics; If the aforementioned video content decoded from elementary streams is determined to be displayable, select a process from the aforementioned additional information and the characteristics of the display and adapt the video content decoded from elementary streams according to the selected process, wherein the aforementioned additional information also indicates whether at least one particular type of metadata is present within the aforementioned signal for a particular duration of the aforementioned elementary streams.
- 6. Processor-readable media CHARACTERIZED in that it comprises a method for adapting video content decoded from elementary streams to the characteristics of a display, as defined in any one of claims 1 to 4.
- 7. Non-transient storage media CHARACTERIZED in that it comprises a method for adapting video content decoded from elementary streams to the characteristics of a display, as defined in any one of claims 1 to 4.
Description
1. Field. [001] This disclosure generally relates to image/video encoding and decoding. Specifically, but not exclusively, the technical scope of this disclosure relates to the delivery of High Dynamic Range (HDR) content on displays with varying capabilities. 2. Fundamentals. [002] This section refers to introducing the reader to various aspects of the technique, which may be related to various aspects of the present principles that are described and/or claimed below. It is believed that this discussion will be useful in providing the reader with background information to facilitate a better understanding of the various aspects of the present principles. Therefore, it should be understood that these statements should be read in that light, and not as admissions of the prior art. [003] Capturing and displaying image content, whether still images or videos, with a dynamic range that approximates real-life scenes, has been a long-term challenge. [004] In fact, human vision has a wide latitude for scene brightness and possesses multiple adaptation mechanisms that provide an automatic gain for the human visual system. Notably, the brightness range that people can see is much greater than the contrast range available from most currently existing displays. [005] In an attempt to offer content display that meets the requirements of human vision, High Dynamic Range (HDR) images have been specified and defined, which include higher peak luminance, lower minimum luminance, and greater contrast range compared to standard dynamic range (SDR) images. In other words, HDR images offer enhanced shadows and highlight details over SDR images, i.e., brighter whites, deeper blacks, and more vivid colors that better match the images we see in the real world. [006] As is well known, an image, whether HDR or SDR, can be represented by one or more sample matrices (also known as pixel values) in a specific encoding format, which specifies all the information relating to those pixel values and all the information that can be used by a viewfinder and/or any other device to decode and/or display this image. The same applies to video content. Hereafter, the term “image content” will be used to cover both images and video content. [007] An image content generally comprises at least one component, in the form of a first sample matrix, usually a luma (or luminance) component, and possibly at least one other component, in the form of at least one other sample matrix, usually a color component. Or, equivalently, the same information can also be represented by a set of color sample matrices, such as the traditional trichromatic RGB representation. [008] A pixel value is represented by a vector of values C, where C is the number of components. Each value in a vector is represented with a number of bits, which defines a maximum dynamic range of pixel values. [009] Broadcasters and distributors of image content are expected to frequently need to convert between SDR and HDR content to support delivery to all possible outputs and devices. This may be the case, notably, for converting HDR content to layered content, which allows backward compatibility with SDR distribution and display systems. It may also be the case for interoperability of HDR signals with HDR displays that have less peak white balance than the source. [010] Displays, image processors, up/down color converters will need to detect the HDR encoding and colorimetry in use to process and display the signal correctly. [011] To facilitate this processing, some Standards Development Organizations (SDOs) have specified a set of metadata information to be associated with HDR (High Dynamic Range) encoded video streams. The purpose of this metadata is to facilitate adaptation (e.g., HDR-to-SDR or HDR-to-HDR remapping) or interpretation of the reconstructed HDR images to current display characteristics. [012] These HDR metadata mainly comprise static metadata, on the one hand, and dynamic or content-dependent metadata, on the other hand, the latter being defined as metadata that can vary dynamically across the source content (i.e., can change by scene within the same sequence). [013] They include, for example: - Mastering Display Color Volume Information (as described, for example, in SMPTE Standard ST 2086:2014, “Mastering Display Color Volume Metadata Supporting High Luminance and Wide Color Gamut Images”, CEA Standard “HDR Static Metadata Extensions CEA-861.3 January 2015 and ITU-T H.265 (10/2014)): this static metadata is included with mastered HDR content to convey the color volume of the mastering display and the luminance of the content. This is described by the chromaticity of the red, green, and blue display primaries, and the white point of the mastering display, as well as its black level and peak luminance level. In other words, this static metadata describes the actual HDR color volume of the mastering display used to create the image content to guide content adaptation (e.g., c