63‐4: A Novel Real‐Time Full‐Color 3D Holographic (Diffractive) Video Capture, Processing, and Transmission Pipeline Using Off‐the‐Shelf Hardware

📅 2022-06-01
🏛️ SID Symposium Digest of Technical Papers
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
This work presents the first real-time, full-color, dynamic 3D holographic video communication system deployed entirely on off-the-shelf commercial hardware, addressing the core bottlenecks of computational intensity and bandwidth constraints in real-time holographic telepresence. Methodologically, we propose an end-to-end pipeline comprising RGBZ acquisition, diffraction computation, lightweight compression, and network transmission: leveraging multi-modal iPhone sensors (RGB + depth) for 3D scene capture, VividQ’s holographic SDK for real-time phase optimization and diffraction pattern generation, and a custom efficient encoding scheme with adaptive transmission protocol. The system achieves 720p@30fps full-color 3D holographic video calls with end-to-end latency under 120 ms, enabling deployment on consumer-grade devices and cross-platform display. Crucially, it requires no custom optical or computational hardware—establishing the first reproducible, scalable pathway toward practical holographic communication.

Technology Category

Application Category

📝 Abstract
This paper details the world's first live 3D holographic (diffractive) video call using off-the-shelf hardware. We introduce a novel pipeline that facilitates the capture, processing, and transmission of RGBZ data, using an iPhone for image and depth capture with VividQ's SDK for hologram generation and hardware for display.
Problem

Research questions and friction points this paper is trying to address.

Real-time 3D holographic video call
Capture, process, transmit RGBZ data
Use off-the-shelf hardware for implementation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Real-time holographic video
Off-the-shelf hardware
RGBZ data processing
🔎 Similar Papers
No similar papers found.
Ankur Samanta
Ankur Samanta
PhD Student, Columbia University
AI ReasoningMulti-modal foundation modelsRLHF
Gregor Mackenzie
Gregor Mackenzie
PhD Student, School of PET, University of York
T
T. Rathkamp
R42 Institute, Palo Alto, CA, USA
A
A. Cable
R42 Institute, Palo Alto, CA, USA
D
D. Milne
VividQ, Cambridge, UK
A
A. Kaczorowski
VividQ, Cambridge, UK
R
Ronjon Nag
R42 Institute, Palo Alto, CA, USA