DIAL: Distribution-Informed Adaptive Learning of Multi-Task Constraints for Safety-Critical Systems

📅 2025-01-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Safety-critical systems (e.g., autonomous driving) suffer from poor generalization and reliance on handcrafted constraint functions in multi-task safe reinforcement learning. Method: We propose the first distribution-aware multi-task constraint adaptation framework, which models cross-task constraint distributions to enable constraint knowledge transfer and dynamically calibrates risk sensitivity via expert demonstrations to mitigate safety generalization bottlenecks induced by expert bias. Technically, it integrates imitation learning, implicit constraint distribution modeling, risk-aware optimization, and multi-task meta-learning. Contribution/Results: Evaluated on control and navigation tasks, our method achieves significant improvements in both safety and task success rate—without requiring any task-specific constraint definitions—outperforming all existing baselines comprehensively.

Technology Category

Application Category

📝 Abstract
Safe reinforcement learning has traditionally relied on predefined constraint functions to ensure safety in complex real-world tasks, such as autonomous driving. However, defining these functions accurately for varied tasks is a persistent challenge. Recent research highlights the potential of leveraging pre-acquired task-agnostic knowledge to enhance both safety and sample efficiency in related tasks. Building on this insight, we propose a novel method to learn shared constraint distributions across multiple tasks. Our approach identifies the shared constraints through imitation learning and then adapts to new tasks by adjusting risk levels within these learned distributions. This adaptability addresses variations in risk sensitivity stemming from expert-specific biases, ensuring consistent adherence to general safety principles even with imperfect demonstrations. Our method can be applied to control and navigation domains, including multi-task and meta-task scenarios, accommodating constraints such as maintaining safe distances or adhering to speed limits. Experimental results validate the efficacy of our approach, demonstrating superior safety performance and success rates compared to baselines, all without requiring task-specific constraint definitions. These findings underscore the versatility and practicality of our method across a wide range of real-world tasks.
Problem

Research questions and friction points this paper is trying to address.

Adaptive Learning
Multi-task Rules
Autonomous Driving Safety
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive Learning
Risk Assessment
Multi-task Safety Rules
🔎 Similar Papers
No similar papers found.
S
Se-Wook Yoo
Department of Electrical and Computer Engineering and ASRI, Seoul National University, Seoul, Republic of Korea
Seung-Woo Seo
Seung-Woo Seo
Seoul National University, Dept of Electrical and Computer Engineering
Reinforcement LearningAutonomous Driving