🤖 AI Summary
Identifying and exploiting unexplored voids—low-density regions in high-dimensional complex data—remains challenging due to their geometric ambiguity and semantic opacity.
Method: We propose a void-driven autonomous discovery paradigm. First, we introduce the Void-Center Search Algorithm (ESA) to locate high-dimensional voids. Second, we design GapMiner, a visualization system integrating linked parallel coordinates with gradient-ascent optimization to enable human-in-the-loop, closed-loop exploration. Third, we develop the Exploration Steering Configuration (ESC) mechanism, enabling deep neural networks to transition from interactive learning to autonomous, gradient-guided void exploration.
Results: Experiments across parameter optimization, adversarial example generation, and reinforcement learning demonstrate that configurations discovered by our method significantly outperform random baselines in novelty and performance. Multiple case studies validate its strong generalizability and practical utility.
📝 Abstract
We present a comprehensive pipeline, augmented by a visual analytics system named ``GapMiner'', that is aimed at exploring and exploiting untapped opportunities within the empty areas of high-dimensional datasets. Our approach begins with an initial dataset and then uses a novel Empty Space Search Algorithm (ESA) to identify the center points of these uncharted voids, which are regarded as reservoirs containing potentially valuable novel configurations. Initially, this process is guided by user interactions facilitated by GapMiner. GapMiner visualizes the Empty Space Configurations (ESC) identified by the search within the context of the data, enabling domain experts to explore and adjust ESCs using a linked parallel-coordinate display. These interactions enhance the dataset and contribute to the iterative training of a connected deep neural network (DNN). As the DNN trains, it gradually assumes the task of identifying high-potential ESCs, diminishing the need for direct user involvement. Ultimately, once the DNN achieves adequate accuracy, it autonomously guides the exploration of optimal configurations by predicting performance and refining configurations, using a combination of gradient ascent and improved empty-space searches. Domain users were actively engaged throughout the development of our system. Our findings demonstrate that our methodology consistently produces substantially superior novel configurations compared to conventional randomization-based methods. We illustrate the effectiveness of our method through several case studies addressing various objectives, including parameter optimization, adversarial learning, and reinforcement learning.