Crossmodal learning for Crop Canopy Trait Estimation

📅 2025-11-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitation of coarse spatial resolution in satellite remote sensing for precise monitoring of crop canopy traits at the micro-parcel level, this paper proposes a novel cross-modal learning framework that directly generates drone-like high-resolution feature representations from low-resolution satellite imagery—achieved for the first time. Our method employs a spectral-spatial jointly aligned deep neural network to explicitly model multi-scale correspondences between satellite and UAV images. Trained on a large-scale, synchronized paired dataset comprising 84 maize varieties across five geographic sites, the model produces high-fidelity features. Experimental results demonstrate substantial improvements in downstream agricultural tasks: yield prediction and nitrogen estimation accuracy increase by an average of 12.7% over baseline satellite inputs. This work effectively bridges the perceptual gap between spaceborne and proximal remote sensing, establishing a scalable, cross-modal representation learning paradigm for precision agriculture.

Technology Category

Application Category

📝 Abstract
Recent advances in plant phenotyping have driven widespread adoption of multi sensor platforms for collecting crop canopy reflectance data. This includes the collection of heterogeneous data across multiple platforms, with Unmanned Aerial Vehicles (UAV) seeing significant usage due to their high performance in crop monitoring, forecasting, and prediction tasks. Similarly, satellite missions have been shown to be effective for agriculturally relevant tasks. In contrast to UAVs, such missions are bound to the limitation of spatial resolution, which hinders their effectiveness for modern farming systems focused on micro-plot management. In this work, we propose a cross modal learning strategy that enriches high-resolution satellite imagery with UAV level visual detail for crop canopy trait estimation. Using a dataset of approximately co registered satellite UAV image pairs collected from replicated plots of 84 hybrid maize varieties across five distinct locations in the U.S. Corn Belt, we train a model that learns fine grained spectral spatial correspondences between sensing modalities. Results show that the generated UAV-like representations from satellite inputs consistently outperform real satellite imagery on multiple downstream tasks, including yield and nitrogen prediction, demonstrating the potential of cross-modal correspondence learning to bridge the gap between satellite and UAV sensing in agricultural monitoring.
Problem

Research questions and friction points this paper is trying to address.

Enhancing satellite imagery resolution with UAV-level detail for crop monitoring
Bridging spatial resolution limitations between satellite and UAV agricultural sensing
Improving crop trait estimation through cross-modal spectral-spatial correspondence learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Crossmodal learning enriches satellite imagery with UAV detail
Model learns spectral-spatial correspondences between sensing modalities
Generated UAV-like representations outperform real satellite imagery
🔎 Similar Papers
No similar papers found.
T
Timilehin T. Ayanlade
Department of Mechanical Engineering, Iowa State University, Ames, IA, USA.
A
Anirudha Powadi
Department of Computer Engineering, Iowa State University, Ames, IA, USA.
T
Talukder Z. Jubery
Department of Mechanical Engineering, Iowa State University, Ames, IA, USA.
Baskar Ganapathysubramanian
Baskar Ganapathysubramanian
NSF/USDA AI Institute, Iowa State University
Computational scienceTransport phenomenaEnergy technologyMachine learning
S
Soumik Sarkar
Department of Mechanical Engineering, Iowa State University, Ames, IA, USA.