Tools
Explore new features and tools within Intel® products, communities, and platforms
77 Discussions

Deliver High-Quality, High-Performance HDR via Intel® oneAPI Video Processing Library

Pamela_H_Intel
Moderator
1 1 2,572

HDR (High Dynamic Range) technology brings life to natural scenes with vivid colors, enables visible content even under challenging lighting conditions, and provides higher contrast in the scenes, thereby allowing better artistic representations. This paper introduces what HDR is and how to deliver HDR in media delivery solutions on Intel® Graphics via Intel® oneAPI Video Processing Library (oneVPL).

Furong Zhang, GPU Software Development Engineer, Intel Corporation

HDR: Unlock New Video Opportunities

With video technology evolving from high-definition (HD) to ultra-high definition (UHD), new video-enhanced characteristics are refining video such as higher resolution, higher frame rate, and higher dynamic range (HDR). A marked improvement over high-definition video technology, HDR is increasing perceived quality, making each pixel able to better represent the full range of brightness we see. This enables the viewer to see a more realistic image and have an immersive visual experience.

Intel oneAPI Video Processing Library (oneVPL) is a public programming interface which delivers fast, high-quality, real-time video decoding, encoding, transcoding, and processing on CPUs, GPUs, and other accelerators for broadcasting, live streaming and VOD, cloud gaming, and more. This paper explains the powerful advantages of HDR and shows how we use oneVPL to deliver these advantages.

HDR Technology Basics

Generally speaking, HDR technology (e.g., HDR video recording and HDR TV display) is a combination of several components including High Dynamic Range, Wide Color Gamut, Higher Bit-Depth (10/12bit sampling), and OETF/EOTF Transfer Functions. With this combination, HDR technology improves the pixels and brings life to natural scenes with vivid colors, enables higher contrast in scenes which allow better artistic representations and enables visible content even under challenging lighting conditions.

High Dynamic Range

The dynamic range of video refers to its luminance range, the maximum and minimum levels. Luminance was measured candela per square meter (cd/m2) – or “nits”. The human eye can adapt to an enormous range of light intensity levels; it can detect a luminance range from 10-6 cd/m2 to 108 cd/m2. The conventional, now referred to as Standard Dynamic Range (SDR) supports luminance values only in the range of 0.01 to 100 cd/m2.  High Dynamic Range is the capability to represent a large luminance variation in the video signal, i.e., from very dark values (less than 0.01 cd/m2) to very bright values (10000 cd/m2). HDR creates brighter whites, darker blacks, and brighter colors that better match images we see in the real world. The enhanced image quality due to HDR is appreciable under virtually any viewing condition.

f1-luminance-in-nits-hdr-vs-sdr.png

Figure 1 – Luminance in Nits HDR VS SDR

Wide Color Gamut

Wide Color Gamut is the capability of representing a wide range of colors than have been supported. The traditional has been based on Rec.709 color space, which only captures a relatively small percentage (35.9%) of all visible chromaticity values, according to CIE 1931. But Wider color spaces, such as Rec.2020 can represent a much larger percentage of visible chromaticities (~75.8 %).

f2-wider-color-gamut-hdr-vs-sdr.png

Figure 2 – Wider Color Gamut HDR vs SDR

Higher Bit-Depth

The conventional uses 8-bit sample precision and does not provide sufficient brightness or color
levels, which could result in the viewer seeing artifacts such as visible banding in the picture. This
limitation is exacerbated with WCG and HDR. Higher sample precision, such as 10-bit or 12-bit, represents more colors and brightness levels and greatly improves the ability to represent smoother transitions between hues or brightness to reduce banding effects.

OETF/EOTF Transfer Function

OETF (Optical Electrical Transfer Function) refers to the way the optical signal gets translated into voltage at the capture (e.g., camera) side and EOTF (Electrical Optical Transfer Function) refers to the way the electrical signal gets translated into optical signal at the display side. These transfer functions are based on gamma, power functions, logarithmic functions, and on perceptual models to mimic the non-linear behavior of the human visual system. Examples of OETF/EOTF transfer functions are Standard Gamma (transfer function for Rec.709/Rec.601), Hybrid Log-Gamma (HLG, developed by BBC & NHK) and Perceptual Quantization (SMPTE ST.2084).

HDR vs SDR Comparison

HDR vs SDR comparison are shown in Table 1

Component

Standard Dynamic Range

High Dynamic Range

Luminance range

Up to 100 nits

Up to 10000 nits

Color Gamut

Rec. 709 in general

DCI-P3, Rec. 2020 in general

Bit Depth

Usually in 8-bits

In 10-bits, 12-bits

Transfer Function

Standard Gamma

PQ ST2084, HLG

Table 1 - HDR vs SDR comparison

HDR Standardization

There is a lot of effort to define new standards for delivering HDR content. The following provides an overview of standards in HDR technology.

SMTPE ST.2084

SMTPE ST.2084: High Dynamic Range Electro-Optical Transfer Function of Mastering Reference Displays

This standard specifies an EOTF characterizing high-dynamic-range reference displays used primarily for mastering non-broadcast content. This standard also specifies an Inverse-EOTF derived from the EOTF.

SMTPE ST.2086

SMTPE ST.2086: Mastering Display Color Volume Metadata Supporting High Luminance and Wide Color Gamut Images

This standard specifies the metadata items to specify the color volume (the color primaries, white
point, and luminance range) of the display that was used in mastering video content. The metadata is
specified as a set of values independent of any specific digital representation.

SMTPE ST.2094

SMTPE ST.2094: Dynamic Metadata for Color Volume Transform

The color transformation process can be optimized through the use of content dependent, dynamic color transform metadata rather than using only display color volume metadata.

HDR Coding

SMPTE ST.2084 and ST.2086 were added to HEVC as part of an extension (SEI: supplemental enhancement information) update in 2014 so it is included in the compressed HEVC file, also the latest AV1.

HDR Types

Actually, there is no final description of HDR types. There are two widely used standards employed today:  HDR10 and Dolby Vision. In addition to these, there is a new HDR10+ format which plans to launch dynamic HDR to the HDR10 standard while staying royalty-free. The BBC of the UK and NHK of Japan have jointly proposed the HLG HDR type which is optimized more for real-time broadcast production and is royalty-free. UVA (UHD Video Alliance in China) also proposed HDR Vivid as the type (mainly for China) with royalty-free. The following table lists a few parameters of HDR types for comparison. 

Types

Transfer Function

Metadata

Max Luminance

HDR10

ST2084

Static

Typically, up to 4000 nits

HDR10+

ST2084

Dynamic

Up to 10000 nits

Dolby Vision

ST2084

Dynamic

Up to 10000 nits

HLG

Hybrid Log Gamma (HLG)

N/A

N/A

HDR Vivid

ST2084 or HLG

Dynamic

Up to 10000 nits

Table 2 – HDR types comparison

Intel® oneAPI Video Processing Library

Overview

Intel® oneAPI Video Processing Library (oneVPL) is a programming interface for video decoding, encoding, and processing to build portable media pipelines on CPUs, GPUs, and other accelerators. The oneVPL API is used to develop quality, performant video applications that can leverage Intel® hardware accelerators. It provides device discovery and selection in media-centric and video analytics workloads, and API primitives for zero-copy buffer sharing. oneVPL is backward-compatible with Intel® Media SDK and cross-architecture compatible to ensure optimal execution on current and next-generation hardware without source code changes. For more details, please visit the official oneVPL website. The following diagram shows how oneVPL is working on Intel® Graphics Stack (Linux*). 

f3-sw-stack-for-media-pipeline-gpu.png

Figure 3 - Intel® software stack to build media pipelines on an Intel® GPU (Linux)

HDR in oneVPL

As the paper introduced in the previous section, HDR technology is the combination of the following components:  High Luminance Dynamic Range, Wide Color Gamut, High bit depth, and EOTF/OEFT transfer function. In the video specification, several supplemental enhancement information (SEI) messages to describe the components of HDR technology are added in HEVC/AV1to support HDR signaling (SEI details, please visit HEVC / AV1 specification). oneVPL defines a few structures which can be mapped to video specification (HEVC/AV1) semantics.

  • API structure mfxExtContentLightLevelInfo, indicate content light level for HDR and can be mapped to video specification(HEVC/AV1) semantics.
  • API structure mfxExtMasteringDisplayColourVolume, indicate mastering display colour volume for HDR and can be mapped to video specification(HEVC/AV1) semantics.
  • API structure mfxExtVideoSignalInfo, indicate video signal info, need to program the correct colour primaries and transfer function for HDR and can be mapped to video specification(HEVC/AV1) semantics.
  • API structure mfxExtVPP3DLut, for HDR Tone mapping (LUT-based).

HDR Transcoding

Here, HDR transcoding means 3 scenarios: HDR input stream-> HDR output stream, HDR input stream-> SDR output stream, SDR input stream-> HDR output stream.

HDR->HDR Transcoding

HDR->HDR transcoding includes HDR decoding and HDR encoding. In this scenario, video processing is optional based on the need (e.g., scaling may be needed). Please see more details about HDR decoding/encoding in the following part.

f4-hdr-to-hdr-transcoding.png

Figure 4 – HDR->HDR transcoding

 

HDR->SDR Transcoding

HDR->SDR transcoding includes HDR decoding, HDR Tone Mapping, and SDR encoding. In this scenario, video processing HDR Tone Mapping is a must-have, other video processing feature is optional based on the need (e.g., scaling may be needed). Please see more details about HDR decoding/tone mapping in the following part.

f5-hdr-to-sdr-transcoding.png

 Figure 5 – HDR->SDR transcoding

 

SDR->HDR Transcoding

SDR->SDR transcoding includes SDR decoding, HDR Tone Mapping, and HDR encoding. In this scenario, video processing HDR Tone Mapping is a must-have, other video processing feature is optional based on the need (e.g., scaling may be needed). Please see more details about HDR encoding/tone mapping in the following part.

f6-sdr-to-hdr-transcoding.png

Figure 6 – SDR->HDR transcoding

HDR Decoding

oneVPL provides HDR decoding functionality. The application should extract HDR SEI information from bitstream via the oneVPL API for HEVC/AV1 HDR decoding. Please visit API structure mfxExtContentLightLevelInfo, mfxExtMasteringDisplayColourVolume, mfxExtVideoSignalInfo for more details.

HDR Encoding

oneVPL provides HDR encoding functionality. The application should add HDR SEI information into bitstream via the oneVPL API for HEVC/AV1 HDR encoding. Please visit API structure mfxExtContentLightLevelInfo, mfxExtMasteringDisplayColourVolume, mfxExtVideoSignalInfo for more details. The application should pay attention to mfxExtVideoSignalInfo, need to set the correct “ColourPrimaries” and “TransferCharacteristics”(for example, Rec.2020 and SMPTE ST.2084).

HDR Tone Mapping

Tone mapping is a technique used in image processing to map one set of colors to another. Correct tone mapping is necessary for a good user experience (please see the comparison in Figure 7a and 7b). Currently, oneVPL provides 2 approaches for HDR Tone Mapping.

f7a-f7b-tone-mapping-comparision.png

Figures 7a & 7b – with correct tone mapping (7a) & without correct tone mapping (7b)

Look Up Table(LUT)- Based HDR Tone Mapping

This approach is flexible; the application can customize the tone mapping effect based on the Look Up table from the application. But this approach requires the application developer has image-processing knowledge and expertise.  Let’s take HLG input->SDR output as an example to explain how it works.

As the HDR Transcoding section introduced, there are 2 typical conversion HDR input signals -> SDR input signal and SDR input signal-> HDR input signal. Take HDR signal p010 format SDR signal nv12 format as the example to explain how to use LUT-based Tone Mapping.

Figure 8 shows the detailed steps of HDR->SDR Tone Mapping. All these steps can be completed on one VPP API call MFXVideoVPP_RunFrameVPPAsync. But the application needs to program the correct input/output (in this case, the input HDR P010 BT2020 input and the output SDR NV12 BT709) and 3DLUT parameters (video-processing-3dlut). In this pipeline, Chroma Upsampling, Color Space Conversion are added automatically when oneVPL builds the data-processing pipeline. The application does not need to take care of it.

f8-hdr-to-sdr-tone-mapping.png

Figure 8 – HDR->SDR Tone Mapping

 

Figure 9 is the detailed steps of SDR->HDR Tone Mapping. Same as HDR->SDR Tone Mapping, the application needs to program the correct input/output (in this case, the input SDR NV12 BT709 and the output HDR P010 BT2020) and 3DLUT parameters (video-processing-3dlut).

 

Figure 9 – SDR->HDR Tone MappingFigure 9 – SDR->HDR Tone Mapping

Figure 9 – SDR->HDR Tone Mapping

 

Regarding 3DLUT-based Tone Mapping, we investigated the 3DLUT mapping accuracy. With 65^3, the deltaE is <2 which can keep the color accuracy.

HDR Metatada-Based HDR Tone Mapping

oneVPL also provides HDR metadata-based Tone Mapping. In this approach, the application only needs to program the HDR metadata as the programming guide HDR Tone Mapping indicates. Currently, this approach only supports HDR10 to SDR.

Sample Code

For HDR transcoding sample code, please visit sample_multi_transcode.

Summary and Conclusion

Intel oneAPI Video Processing Library provides the open programming interface for fast, high-quality, real-time HDR decoding, tone mapping, and encoding which delivers the HDR advantages. With oneVPL, the application developer can develop quality, performant video applications that can leverage Intel hardware accelerators.

Bibliography/References

  1. oneVPL specification: https://spec.oneapi.io/versions/latest/elements/oneVPL/source/index.html
  2. Introduction to deltaE: https://www.stellarnet.us/delta-e-%CE%B4e-used-colorimetry/
  3. CIE system: http://www.colorbasics.com/CIESystem/
  4. SMTPE ST.2084, SMTPE ST.2086, SMTPE ST2094

Legal Information

No license (express or implied, by estoppel or otherwise) to any intellectual property rights is granted by this document.

Intel disclaims all express and implied warranties, including without limitation, the implied warranties of merchantability, fitness for a particular purpose, and non-infringement, as well as any warranty arising from course of performance, course of dealing, or usage in trade.

This document contains information on products, services and/or processes in development.  All information provided here is subject to change without notice. Contact your Intel representative to obtain the latest forecast, schedule, specifications, and roadmaps.

The products described in this document may contain design defects or errors known as errata which may cause the product to deviate from published specifications. Current characterized errata are available on request.

See Related Content

Tech Articles

  • oneVPL Support Added to FFMPEG for Great Streaming on Intel® GPUs 
    Read

On-Demand Webinars

  • Optimize and Accelerate Video Pipelines across Architectures
    Watch
  • Accelerate Video Processing on More CPUs, GPUs, and Accelerators
    Watch
About the Author
Software engineer for 20+ years. Excels in all things software, plus connecting people and teams for optimal synergy.
1 Comment
Furong_Zhang_Intel

Looks great to me.