Knowledge Grafting: A Mechanism for Optimizing AI Model Deployment in Resource-Constrained Environments

📅 2025-07-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of deploying large language models under resource-constrained conditions, this paper proposes a novel knowledge transfer paradigm termed “knowledge grafting”—an analogy to horticultural grafting—that selectively transfers critical features from a large teacher model to a lightweight student model. The method integrates knowledge distillation, adaptive feature selection, parameter pruning, and collaborative fine-tuning to construct a high-performance rootstock model. Experimental results demonstrate that the proposed approach compresses the model size from 64.39 MB to 7.38 MB (an 88.5% reduction), while improving validation accuracy to 89.97% and test accuracy to 90.45%, alongside a significant reduction in validation loss. Compared with conventional model compression techniques, our method simultaneously maintains low computational overhead and overcomes the performance bottlenecks of small models in terms of generalization capability and predictive accuracy. This work establishes a new paradigm for efficient edge AI deployment.

Technology Category

Application Category

📝 Abstract
The increasing adoption of Artificial Intelligence (AI) has led to larger, more complex models with numerous parameters that require substantial computing power -- resources often unavailable in many real-world application scenarios. Our paper addresses this challenge by introducing knowledge grafting, a novel mechanism that optimizes AI models for resource-constrained environments by transferring selected features (the scion) from a large donor model to a smaller rootstock model. The approach achieves an 88.54% reduction in model size (from 64.39 MB to 7.38 MB), while improving generalization capability of the model. Our new rootstock model achieves 89.97% validation accuracy (vs. donor's 87.47%), maintains lower validation loss (0.2976 vs. 0.5068), and performs exceptionally well on unseen test data with 90.45% accuracy. It addresses the typical size vs performance trade-off, and enables deployment of AI frameworks on resource-constrained devices with enhanced performance. We have tested our approach on an agricultural weed detection scenario, however, it can be extended across various edge computing scenarios, potentially accelerating AI adoption in areas with limited hardware/software support -- by mirroring in a similar manner the horticultural grafting enables productive cultivation in challenging agri-based environments.
Problem

Research questions and friction points this paper is trying to address.

Optimizes AI models for resource-constrained environments
Reduces model size while improving generalization capability
Enables AI deployment on devices with limited resources
Innovation

Methods, ideas, or system contributions that make the work stand out.

Knowledge grafting transfers features to smaller models
Reduces model size by 88.54% while improving accuracy
Enables AI deployment in resource-constrained edge devices
🔎 Similar Papers
No similar papers found.
O
Osama Almurshed
Prince Sattam Bin Abdulaziz University, Kingdom of Saudi Arabia
A
Ashish Kaushal
Indian Institute of Technology Ropar, India
A
Asmail Muftah
Azzaytuna University, Libya
Nitin Auluck
Nitin Auluck
Computer Science Indian Institute of Technology Ropar
Real-Time SystemsSchedulingEdge Computing.
Omer Rana
Omer Rana
Cardiff University, UK
High Performance Distributed ComputingEdge ComputingData Analysis and MiningMulti-Agent