🤖 AI Summary
To address bandwidth limitations in immersive remote visualization—causing degraded resolution and fidelity—this paper proposes an eye-tracking-driven foveated compression method. It integrates real-time eye-tracking data with standard video encoders (e.g., H.264/AVC) to dynamically modulate spatial-domain quantization parameters: preserving high fidelity in the foveal region while substantially reducing bitrates in peripheral areas, thereby achieving retinal-perception-aligned non-uniform compression. We design a lightweight, plug-and-play foveated compression framework and present its first end-to-end validation on the award-winning NimbRo Avatar system. Experiments demonstrate that, without compromising subjective immersion, average bandwidth consumption drops to 33% of the original stream. The approach thus achieves an optimal trade-off between compression efficiency and visual quality, establishing a novel paradigm for bandwidth-constrained telepresence and teleoperation applications.
📝 Abstract
Immersive televisualization is important both for telepresence and teleoperation, but resolution and fidelity are often limited by communication bandwidth constraints. We propose a lightweight method for foveated compression of immersive televisualization video streams that can be easily integrated with common video codecs, reducing the required bandwidth if eye tracking data is available. Specifically, we show how to spatially adjust the Quantization Parameter of modern block-based video codecs in a adaptive way based on eye tracking information. The foveal region is transmitted with high fidelity while quality is reduced in the peripheral region, saving bandwidth. We integrate our method with the NimbRo avatar system, which won the ANA Avatar XPRIZE competition. Our experiments show that bandwidth can be reduced to a third without sacrificing immersion. We analyze transmission fidelity with qualitative examples and report quantitative results.