Building Blocks for 3D Displays: A New Hologram Dataset Fuels Machine Learning

Author: Denis Avetisyan


Researchers have released a large-scale dataset designed to accelerate the development of high-resolution, large-depth-range 3D displays powered by machine learning.

Holographic reconstructions demonstrate focused capabilities across varying depths, with methods achieving clarity at both near and far planes despite inherent limitations in practical application.
Holographic reconstructions demonstrate focused capabilities across varying depths, with methods achieving clarity at both near and far planes despite inherent limitations in practical application.

KOREATECH-CGH provides a comprehensive resource for training and evaluating algorithms in layer-based computer-generated holography.

Despite recent advances in machine learning-based computer-generated holography (ML-CGH), a key limitation remains the scarcity of large-scale, high-quality datasets for training robust 3D reconstruction models. This work introduces KOREATECH-CGH, a publicly available dataset comprising 6,000 RGB-D image pairs and corresponding complex holograms-spanning resolutions up to 2048×2048 and extended depth ranges-along with a novel amplitude projection technique to enhance reconstruction fidelity. By demonstrating improved performance over existing layer-based methods and validating its utility through hologram generation and super-resolution tasks, KOREATECH-CGH offers a valuable resource for accelerating the development of next-generation 3D displays-but how will this dataset enable entirely new applications of ML-CGH beyond conventional imaging?


Beyond the Illusion: Why Realistic 3D Still Feels Out of Reach

The creation of three-dimensional models has historically been a laborious process, demanding significant artistic skill and countless hours of manual refinement. While digital tools have streamlined aspects of the workflow, achieving true photorealism – the convincing replication of light and material properties – remains a substantial challenge. This limitation directly impacts the burgeoning fields of augmented and virtual reality, where immersive experiences rely on convincingly rendered objects and environments. Beyond entertainment, industries like product design, architectural visualization, and medical imaging require increasingly accurate and detailed 3D representations, and the inability to rapidly generate these models with sufficient fidelity currently constrains innovation and widespread adoption. The demand for more efficient and realistic 3D modeling techniques is therefore not merely aesthetic, but a fundamental requirement for unlocking the full potential of these technologies.

The pursuit of truly convincing holographic displays faces significant hurdles tied to both the sheer computational demands and the intricacies of rendering light fields. Existing rendering techniques often struggle to accurately simulate the wave-like properties of light necessary for realistic depth and parallax, requiring vastly more processing power than currently available for real-time applications. Moreover, accurately capturing and reconstructing a 3D scene as a hologram necessitates solving complex diffraction calculations for every point in the displayed image, a task that quickly overwhelms even high-end processors. Researchers are actively exploring novel algorithms – including those leveraging machine learning – and specialized hardware architectures to accelerate these computations and overcome the limitations of traditional rendering pipelines, ultimately aiming to deliver holographic experiences that are visually indistinguishable from reality.

The pursuit of instantly generated, lifelike 3D representations is rapidly reshaping both how images are captured and how they are displayed. This demand isn’t simply about sharper visuals; it necessitates breakthroughs in computational imaging, where software algorithms actively participate in the image formation process, extracting depth and detail beyond the capabilities of traditional optics. Researchers are developing novel sensor arrays and light field cameras, coupled with advanced algorithms-including those leveraging AI and neural radiance fields-to reconstruct scenes with unprecedented fidelity in real-time. Simultaneously, display technologies are evolving to accommodate these complex datasets, pushing the boundaries of holographic and volumetric displays to convincingly render these reconstructed 3D worlds, and promising immersive experiences for applications ranging from remote collaboration to medical visualization and beyond.

Optical reconstructions demonstrate focused holograms achieved at both front and back planes, indicating successful depth control.
Optical reconstructions demonstrate focused holograms achieved at both front and back planes, indicating successful depth control.

The Machine Learning Patch: Trading Calculations for Data

Traditional Computer Generated Holography (CGH) relies on computationally intensive rendering processes, specifically the diffraction calculation from 3D scenes, which become significant bottlenecks for real-time applications and high resolutions. Machine learning techniques, however, present an alternative by learning the direct mapping between input 3D scene data and the resulting complex holographic wavefront. This data-driven approach bypasses the need for explicit diffraction calculations, offering substantial speedups and enabling the generation of holograms with reduced computational cost. The efficacy of this method is predicated on the availability of large datasets to train the machine learning models, allowing them to accurately approximate the complex relationship between 3D scenes and their holographic representations.

U-Net and Swin-Unet architectures facilitate direct computation of complex wavefronts for Computer Generated Holography (CGH) through supervised learning. These deep learning models ingest input data, typically in the form of RGB-D images representing color and depth information, and are trained to predict the corresponding complex-valued hologram. This approach bypasses traditional physics-based simulation methods, which are computationally expensive. The models learn the mapping between 3D scene representation and holographic interference patterns, effectively functioning as a data-driven approximation of the wave propagation process. Training requires paired datasets of 3D scene data and the desired complex holograms, enabling the network to optimize its internal parameters to minimize the reconstruction error.

The KOREATECH-CGH dataset is a critical resource for the development of machine learning algorithms applied to computer generated holography (CGH). It comprises 6,000 paired examples of RGB-D data and corresponding complex holograms, enabling supervised training of models to directly reconstruct holographic wavefronts from 3D scene information. The dataset’s high resolution, with images reaching 2048×2048 pixels, is particularly significant as it allows for the training of algorithms capable of generating high-fidelity holograms with fine details. The availability of this large, paired dataset addresses a key limitation in the field, which previously lacked sufficient data to effectively train deep learning models for CGH reconstruction.

Machine learning-generated computer holograms, created from RGB-D images using models like TensorHolography, U-Net, and Swin-Unet, successfully reconstruct focal planes ranging from -1.2 mm to -13 mm.
Machine learning-generated computer holograms, created from RGB-D images using models like TensorHolography, U-Net, and Swin-Unet, successfully reconstruct focal planes ranging from -1.2 mm to -13 mm.

Chasing Real-Time: The Hardware Bottleneck

Generating digital holograms is computationally intensive due to the need to calculate the interference pattern of a large number of wavefronts, typically on the order of 10^6 to 10^9 samples, to represent the 3D scene. This calculation involves complex arithmetic operations for each sample, and the required processing scales quadratically with image resolution. Conventional CPUs struggle to meet the real-time demands of holographic displays, particularly for dynamic scenes or high-resolution imagery, due to limitations in parallel processing capabilities and memory bandwidth. The sheer volume of data and the complexity of the calculations necessitate specialized hardware and algorithmic optimizations to achieve practical holographic reconstruction rates.

Real-time holographic rendering necessitates specialized hardware acceleration due to the computational demands of simulating light wave interference and diffraction. Graphics Processing Units (GPUs) offer massively parallel processing capabilities suitable for the numerous calculations involved in hologram generation. Field-Programmable Gate Arrays (FPGAs) provide reconfigurable hardware architectures allowing for custom acceleration of specific holographic algorithms, offering a balance between performance and flexibility. Application-Specific Integrated Circuits (ASICs) represent the highest level of hardware acceleration, delivering peak performance for dedicated holographic tasks but with limited adaptability. The selection of hardware depends on the specific application requirements, balancing cost, power consumption, and the need for real-time performance.

The Angular Spectrum Method (ASM) and Fresnel Propagation are computationally intensive algorithms used to simulate the propagation of wavefronts, critical for holographic reconstruction. Optimization for parallel processing architectures, such as GPUs, FPGAs, and ASICs, involves techniques like pre-computation of Fourier transforms and efficient memory access patterns. Specifically, ASM relies heavily on Fourier transforms to move between the frequency and spatial domains, making it well-suited for GPU acceleration. Fresnel Propagation, while potentially requiring fewer computations, benefits from the parallel nature of these architectures to handle the large number of sample points required for accurate simulation. These optimizations reduce the computational complexity from O(N^2) to more manageable levels, where N represents the number of sample points, enabling real-time or near real-time holographic rendering.

Different computational methods produce distinct amplitude distributions in generated complex holograms.
Different computational methods produce distinct amplitude distributions in generated complex holograms.

Fine-Tuning the Illusion: Correction and Validation

Layer-Based Holography (LBH) constructs three-dimensional scenes by dividing the target object into discrete layers, enabling precise control over light field manipulation. This method, when coupled with Amplitude Projection (AP), effectively manages the amplitude of light waves, crucial for generating bright and high-contrast holographic images. The AP-LBM (Amplitude Projection – Layer-Based Method) further refines this process by incorporating a computational technique that optimizes light propagation through each layer, minimizing interference and maximizing image quality. By systematically addressing each layer, LBH with AP-LBM facilitates the creation of holograms with improved resolution and reduced artifacts compared to single-layer approaches, allowing for a more structured and controlled rendering of complex 3D scenes.

Phase optimization techniques address distortions inherent in holographic reconstruction by modulating the phase of the light field used to generate the hologram. These techniques iteratively adjust the calculated interference pattern to minimize discrepancies between the desired 3D object and the reconstructed wavefront. Common algorithms employed include stochastic parallel gradient descent and iterative Fourier transform techniques, which refine the hologram’s phase distribution to reduce artifacts like speckle noise and improve image sharpness. This process effectively corrects for aberrations introduced during hologram creation or reconstruction, resulting in a more accurate and visually clear 3D representation of the original object.

Holographic reconstruction quality is quantitatively assessed using Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM). Reported values of PSNR = 27.01 and SSIM = 0.87 demonstrate a high degree of fidelity between the original and reconstructed images. These metrics were obtained over a measured depth of field of 20.334 mm, indicating consistent image quality across this range. Validation is further supported by the implementation of Focal Image Projection, a technique used to verify the accurate reproduction of three-dimensional information within the holographic reconstruction.

Holographic projections of a target image were successfully generated using several lensless computational imaging methods-specifically, SM-LBM, ADV-LBM, and AP-LBM-as demonstrated by the focal image projections.
Holographic projections of a target image were successfully generated using several lensless computational imaging methods-specifically, SM-LBM, ADV-LBM, and AP-LBM-as demonstrated by the focal image projections.

Beyond the Screen: The Long Road to Practical Holography

The advent of real-time, high-fidelity holographic displays represents a significant leap toward truly immersive augmented and virtual reality experiences. Beyond the limitations of traditional two-dimensional screens, these displays generate three-dimensional images that appear to float in space, interacting with light and perspective as if they were physically present. This capability extends far beyond entertainment, with potential applications reshaping education through interactive anatomical models and remote collaboration via life-sized holographic presence. Furthermore, advancements promise to revolutionize communication, allowing for more engaging and realistic teleconferencing and potentially even holographic messaging. The technology’s promise lies in its capacity to blur the lines between the digital and physical worlds, fostering a deeper sense of presence and facilitating novel forms of interaction that were previously confined to the realm of science fiction.

The evolution of three-dimensional visualization is increasingly reliant on a synergistic approach combining sophisticated algorithms with powerful computing resources. Current advancements aren’t simply about increasing processing speed; they involve machine learning techniques that allow systems to learn how to reconstruct realistic holographic images from limited data. These data-driven algorithms refine the process of light field rendering and diffraction pattern calculation, optimizing for both fidelity and computational efficiency. Hardware acceleration, particularly through the use of specialized processors and graphics cards, further enhances this capability, enabling real-time holographic reconstructions that were previously unattainable. This convergence of algorithmic innovation and computational power is not merely improving existing holographic technology, but fundamentally expanding the possibilities for immersive experiences and complex data visualization.

Advancements in holographic reconstruction are increasingly reliant on the quality and scope of training datasets, and the KOREATECH-CGH dataset represents a significant leap forward in this regard. Unlike its predecessor, the MIT-CGH-4K dataset which was limited to a reconstruction depth of just 6 millimeters, KOREATECH-CGH allows for substantially more realistic holographic projections, extending that range to 80 millimeters. This expanded depth capability is crucial for creating truly immersive three-dimensional experiences, as it allows for the accurate portrayal of objects with greater volume and complexity. Consequently, algorithms trained on KOREATECH-CGH can generate holograms that appear more convincingly solid and lifelike, paving the way for practical applications in fields like augmented reality, virtual reality, and advanced display technologies.

The KOREATECH-CGH system generates holograms by rendering RGB and depth maps using OptiX, and then calculating amplitude and phase distributions via the AP-LBM method.
The KOREATECH-CGH system generates holograms by rendering RGB and depth maps using OptiX, and then calculating amplitude and phase distributions via the AP-LBM method.

The pursuit of increasingly complex algorithms for computer-generated holography, as demonstrated by the KOREATECH-CGH dataset, inevitably introduces a new class of fragility. It’s a familiar pattern; each optimization, each layer added to achieve greater depth range and resolution, creates a more elaborate surface for entropy to exploit. Geoffrey Hinton once observed, “Everything optimized will one day be optimized back.” This sentiment rings particularly true when considering machine learning models applied to complex physical phenomena. The dataset itself is merely a snapshot; production environments, with their unpredictable inputs and edge cases, will relentlessly reveal the limitations of even the most meticulously crafted algorithms, demanding constant refinement and, ultimately, a return to simpler, more robust solutions. The elegance of a theoretical framework is always tempered by the reality of deployment.

The Road Ahead

The proliferation of datasets rarely solves fundamental problems; it merely shifts the bottlenecks. KOREATECH-CGH offers a wider testing ground, certainly, but the core challenge remains: turning simulated light fields into something that doesn’t look like a blurry approximation of reality. The bug tracker will fill with new, more subtle artifacts, and the metrics will need constant recalibration as the resolution arms race continues. The claim of ‘large depth range’ invites scrutiny; any system will eventually encounter the limits of coherence and the tyranny of speckle.

Future work will inevitably focus on the ‘end-to-end’ illusion. The dataset provides training data, but it doesn’t address the intractable issues of optical aberrations, display non-uniformities, or the human visual system’s relentless ability to detect even the smallest inconsistencies. Expect to see a divergence between research focused on generating more data and research focused on generating better light. The latter will be far less glamorous, and likely underfunded.

It’s a reasonable step, this curation of holographic scenes. But a dataset is not a solution; it’s an invitation to a more complex failure state. The problem isn’t a lack of data; it’s the inherent difficulty of reconstructing a wavefront. It’s not deployment; it’s letting go.


Original article: https://arxiv.org/pdf/2512.21040.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2025-12-27 17:12