Robust Transfer Learning with Side Information

📅 2026-03-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of excessive conservatism in conventional distributionally robust optimization (DRO) when environments undergo significant shifts, which often leads to overly conservative policies and poor transferability. To mitigate this, the authors propose a novel approach that integrates a small number of samples from the target domain with auxiliary dynamic information—such as moment bounds, distributional distances, and density ratios—between source and target domains. By constraining the transition kernel estimation, the method constructs a tight, estimation-centered uncertainty set that preserves robustness while substantially reducing policy suboptimality and improving sample efficiency. Theoretical analysis bridges DRO, constrained kernel estimation, and finite-sample guarantees. Empirical evaluations on OpenAI Gym and classical control tasks consistently outperform both robust and non-robust transfer learning baselines.

Technology Category

Application Category

📝 Abstract
Robust Markov Decision Processes (MDPs) address environmental shift through distributionally robust optimization (DRO) by finding an optimal worst-case policy within an uncertainty set of transition kernels. However, standard DRO approaches require enlarging the uncertainty set under large shifts, which leads to overly conservative and pessimistic policies. In this paper, we propose a framework for transfer under environment shift that derives a robust target-domain policy via estimate-centered uncertainty sets, constructed through constrained estimation that integrates limited target samples with side information about the source-target dynamics. The side information includes bounds on feature moments, distributional distances, and density ratios, yielding improved kernel estimates and tighter uncertainty sets. The side information includes bounds on feature moments, distributional distances, and density ratios, yielding improved kernel estimates and tighter uncertainty sets. Error bounds and convergence results are established for both robust and non-robust value functions. Moreover, we provide a finite-sample guarantee on the learned robust policy and analyze the robust sub-optimality gap. Under mild low-dimensional structure on the transition model, the side information reduces this gap and improves sample efficiency. We assess the performance of our approach across OpenAI Gym environments and classic control problems, consistently demonstrating superior target-domain performance over state-of-the-art robust and non-robust baselines.
Problem

Research questions and friction points this paper is trying to address.

Robust Transfer Learning
Environment Shift
Distributionally Robust Optimization
Uncertainty Set
Markov Decision Processes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Robust Transfer Learning
Side Information
Estimate-Centered Uncertainty Sets
Distributionally Robust Optimization
Markov Decision Processes
🔎 Similar Papers
No similar papers found.
A
Akram S. Awad
Department of Electrical and Computer Engineering, University of Central Florida, USA
Shihab Ahmed
Shihab Ahmed
PhD Student at Towson University
System SecurityMalware AnalysisNatural Language ProcessingMachine Learning
Yue Wang
Yue Wang
University of Central Florida
Reinforcement LearningOptimizationgame theory
G
George K. Atia
Department of Electrical and Computer Engineering, University of Central Florida, USA; Department of Computer Science, University of Central Florida, USA