๐ค AI Summary
Structured pruning of multi-component neural architectures (MCNAs) under resource constraints often compromises functional integrity due to indiscriminate removal of inter- and intra-component connections. Method: We propose a component-aware pruning framework that constructs a fine-grained dependency graph distinguishing intra- and inter-component data flows, enabling the first instance of localized structured pruning aligned with component functional boundaries. Our approach integrates parameter-group sensitivity analysis with control-task-driven sparsity validation to preserve critical execution paths. Contribution/Results: Experiments on representative MCNAs demonstrate a substantial average pruning rate improvement of +23.6%, while limiting accuracy degradation to within 1.2%. The method thus enables efficient deployment of complex deep neural networks on edge devices without sacrificing functional correctness or task performance.
๐ Abstract
Deep neural networks (DNNs) deliver outstanding performance, but their complexity often prohibits deployment in resource-constrained settings. Comprehensive structured pruning frameworks based on parameter dependency analysis reduce model size with specific regard to computational performance. When applying them to Multi-Component Neural Architectures (MCNAs), they risk network integrity by removing large parameter groups. We introduce a component-aware pruning strategy, extending dependency graphs to isolate individual components and inter-component flows. This creates smaller, targeted pruning groups that conserve functional integrity. Demonstrated effectively on a control task, our approach achieves greater sparsity and reduced performance degradation, opening a path for optimizing complex, multi-component DNNs efficiently.