Table Of Content
With advances in display technology, High Dynamic Range (HDR) content has become the visual standard due to its wide luminance and rich colors. However, Standard Dynamic Range (SDR) content lacks sufficient brightness and color to fully meet HDR display requirements.
To solve this, UniFab developed an end-to-end SDR-to-HDR conversion method using deep learning that combines convolutional and generative adversarial networks. Through multi-task loss, adaptive tone mapping, and gamut expansion, it enhances details and color, improving dynamic range and realism.
The method also integrates supervised and unsupervised training with diverse data augmentation to boost model generalization and adaptability, offering a solid solution for high-quality SDR-to-HDR conversion across various applications.
Dynamic range refers to the luminance ratio between the brightest highlights and the darkest shadows in a video signal, reflecting the richness of details in light and dark areas and the visual expressiveness of the image. The larger the dynamic range, the more abundant the details in light and shadow the video can display, thereby enhancing realism and immersive experience.
SDR
HDR
Compared to SDR, HDR not only expands the numeric range between the brightest and darkest points but also provides finer detail in midtones and shadow areas. This performance is limited by the peak brightness and contrast capabilities of the display device. Furthermore, different HDR formats (such as HDR10, Dolby Vision, and HLG) vary in color gamut, dynamic metadata, and color depth, affecting the final visual effect.
If you watch HDR videos on a non-HDR display, you may see the entire image appear grayish, with lower contrast and saturation, as shown in the example above.
The above image was captured using an HDR display👆.
SDR (Standard Dynamic Range) content typically lacks detail in highlights and shadows, as well as a wide range of color information. HDR (High Dynamic Range) requires the image to have a much greater brightness range and a broader color gamut. Simple linear contrast stretching can result in image distortion, color shifts, or increased noise. Therefore, AI-driven SDR to HDR conversion must ensure naturalness and authenticity of the image while generating new visual details.
UniFab has developed a multimodal deep learning pipeline for high-fidelity SDR to HDR mapping. The core model combines CNNs and GANs to perform content-adaptive feature extraction and style transfer, accurately restoring luminance gradients and color distributions from limited dynamic range inputs, resulting in HDR images with authentic details and natural perception.
2.1.1 Hierarchical Convolutional Neural Network (CNN)
UniFab uses a deep multi-scale convolutional network designed for SDR images. It combines local spatial details with global semantic context via cross-layer fusion and skip connections, allowing more precise reconstruction of luminance and color in HDR restoration.
2.1.2 Conditional Generative Adversarial Network (Conditional GAN)
UniFab, based on the conditional GAN framework and tailored to the data characteristics of SDR images, constructs an end-to-end mapping model to accurately expand the limited SDR color gamut and dynamic range into the color and luminance space of High Dynamic Range (HDR) images.
UniFab employs a multi-task loss framework to improve SDR-to-HDR conversion by jointly optimizing pixel accuracy, structural consistency, visual quality, and color fidelity. The key loss components are:
Multi-task Loss Function
To achieve high-quality SDR to HDR conversion, UniFab has designed and implemented a systematic brightness and color dynamic expansion algorithm. This algorithm ensures the enhancement of HDR images in terms of color richness and visual realism through multi-level and multi-dimensional strategies.
Adaptive Tone Mapping
This module uses content-aware algorithms to dynamically adjust brightness and contrast for HDR output. It balances local and global contrast through multi-scale adjustments and spatial filters, ensuring natural brightness transitions without detail loss. Brightness mapping adapts based on image statistics and scene lighting, smoothly expanding SDR’s limited range to HDR’s wider range, preserving realism and preventing visual fatigue. Saturation is also adaptively controlled to avoid color oversaturation and distortion.
Color Gamut Expansion
SDR content generally uses Rec.709, while HDR uses the wider Rec.2020 color gamut. UniFab’s module employs deep learning-based color mapping to restore colors lost due to SDR’s gamut limits. Accurate color space transformations ensure minimal distortion, while manual settings allow professional fine-tuning of tone, saturation, and hue for customized looks.
Local Contrast Enhancement
Multi-scale techniques enhance contrast locally and globally by decomposing brightness layers via Laplacian pyramid or wavelets to avoid artifacts. Edge and texture details are preserved through guided filtering and edge detection, while adaptive contrast adjustment maintains shadow detail and prevents highlight overexposure, achieving balanced, rich contrast overall.
To ensure UniFab’s generalization and accuracy in SDR to HDR conversion, we designed a comprehensive dataset construction and diversified training strategy, covering supervised, unsupervised, and data augmentation techniques for various complex environments and scenarios.
Paired Supervised Training: Using high-quality, scene-rich paired SDR-HDR images, the model is trained under supervision to learn dynamic range expansion, detail restoration, and color mapping. Paired data includes professionally captured and post-processed HDR synthesized images, ensuring precise matching and authenticity.
Unsupervised Learning: Incorporating CycleGAN and related unsupervised frameworks to train without paired HDR references. This approach uses cycle-consistency loss to ensure reversible mappings and domain adaptation to mitigate distribution discrepancies, enhancing data diversity and generalization—ideal for applications like surveillance and live streaming.
Data Augmentation
UniFab HDR Upconverter AI integrates four deep learning models tailored for different video resolutions. It effectively addresses challenges in dynamic range expansion, color mapping, and detail restoration during SDR to HDR conversion, ensuring superior image quality and natural visual transitions across various scenes and resolutions.
Fast Model
FHD (1080p) Model
QHD Model
4K UHD Model
UniFab HDR Upconverter AI supports the two leading HDR color space standards—HDR DCI-P3 and HDR Rec.2020—allowing users to choose the optimal output based on their display devices and application needs. Leveraging advanced deep learning and image processing techniques, it achieves precise mapping from limited dynamic range and gamut SDR images to wide HDR color spaces, with key features as follows:
1. Color Space Mapping and Conversion
Adaptive color transformation network: Uses conditional GAN with multi-scale convolutions to dynamically learn nonlinear mappings from SDR to target HDR spaces (DCI-P3 or Rec.2020), enabling fine control over high-dimensional color distributions.
Multi-component color correction: Applies linear and nonlinear calibrations per RGB channel through color matrix transformations and gamma correction to ensure accurate and consistent colors.
Joint gamma and tone mapping optimization: Combines HDR-specific transfer functions (PQ or HLG) to enhance shadow details and highlight color fidelity, preventing distortion and color shifts.
2. Dynamic Range Expansion and Luminance Mapping
Multi-scale luminance modeling: Extracts low-frequency brightness and high-frequency textures via deep convolutional features, supporting detail enhancement and global luminance extension.
Adaptive highlight and shadow enhancement: Uses local contrast enhancement and nonlinear brightness stretch to restore bright details and reveal shadow information, enriching HDR luminance and smooth transitions.
Dynamic tone mapping: Adjusts tone curves in real-time based on scene content to avoid brightness compression and highlight clipping during SDR-to-HDR conversion, improving visual comfort.
3. Color Fidelity and Visual Consistency
High-dimensional color distribution matching: Jointly trains with perceptual and color consistency losses to ensure high fidelity in tone, saturation, and contrast, minimizing visual deviations.
Non-local self-attention: Captures long-range pixel color correlations to prevent color discontinuities and artifacts, enhancing natural color gradation and overall image cohesion.
To meet real-time HDR conversion demands for high-resolution videos, UniFab optimized the architecture and operator-level acceleration of the 3027 version model, significantly boosting inference speed. This ensures efficient SDR-to-HDR mapping while maintaining high-quality HDR visuals. Below are the measured conversion performance data and key technical measures based on a 9-minute 57-second, 24fps test clip across various GPU platforms:
Test Source Duration: 9min 57s Frame Rate: 24fps | |||
GPU | Model | Test Results | |
NVIDIA GeForce RTX 4060 Ti | 4K UHD Model | UniFab 3027 Conversion time 2h 46m 40s Speed 1.44fps | UniFab 3028 Conversion time 1h 3m 32s Speed 3.78fps |
NVIDIA GeForce RTX 5070 | QHD Model | UniFab 3027 Conversion time 30m 16s Speed 7.71fps | UniFab 3028 Conversion time 27m 48s Speed 8.45fps |
NVIDIA GeForce RTX 3050 | FHD (1080p) Model | UniFab 3027 Conversion time 51m 8s Speed 4.93fps | UniFab 3028 Conversion time 26m 9s Speed 9.04fps |
NVIDIA GeForce RTX 2080 | Fast Model | UniFab 3027 Conversion time 15m Speed 16-17fps | UniFab 3028 Conversion time 10m Speed stable 16-25fps |
Technical Details and Optimization
Deeper Network Optimization and Automated Search
We plan to introduce Neural Architecture Search (NAS) to automatically explore and design more efficient network structures, further improving SDR-to-HDR conversion quality and inference speed while balancing model compactness and performance.
Multimodal Sensor Data Integration
We aim to incorporate auxiliary data such as sensor depth information and scene lighting measurements to build multimodal fusion models, significantly enhancing dark detail recovery and HDR mapping accuracy under complex lighting.
Continuous Improvement of Multi-task Loss Functions
We plan to adopt loss functions aligned with human visual perception, such as adaptive loss weighting based on visual models and guided by subjective user feedback, to better match the perceived realism.
Support for Multiple HDR Formats
Future releases will support additional HDR standards (e.g., HDR10+, Dolby Vision dynamic metadata), enabling broader device compatibility and content ecosystem integration, enhancing system versatility and market competitiveness.
Cloud Deployment and Edge Computing Optimization
UniFab will keep strengthening its technological edge and application scope in SDR-to-HDR conversion, delivering higher quality, efficiency, and wider coverage HDR visual solutions to meet evolving user needs and diverse scenarios.
👉 Community Link:
If you have topics of interest or models to follow, please leave a message in the forum. We will seriously consider your testing and evaluation suggestions and regularly publish professional technical reviews and upgrade reports.
Next Article Preview: UniFab HDR Upconverter AI|Technical Principle
Previous Articles:UniFab Smoother AI
📗Unifab Colorizer AI: Working Principles and Evolution Through Iterations
📘The Iterations of UniFab Face Enhancer AI
📙UniFab Texture Enhanced: Technical Analysis and Real-World Data Comparison