🤖 AI Summary
Original Neural Fitted Q-Iteration (NFQ) suffers from poor reproducibility, intricate hyperparameter tuning, and limited robustness in real-world control tasks. To address these limitations, this paper proposes NFQ2.0—a refined variant incorporating batch learning to enable seamless transition between online and offline training, a deep multilayer neural network architecture, systematic hyperparameter optimization, and ablation studies to isolate the impact of key design choices. On the CartPole benchmark, NFQ2.0 significantly improves training stability and consistency of policy performance. Furthermore, its engineering viability is validated on a physical testbed built from real industrial components. This work delivers a lightweight, robust, and easily tunable algorithmic framework for reproducible deployment of deep reinforcement learning in industrial control applications.
📝 Abstract
This article revisits the 20-year-old neural fitted Q-iteration (NFQ) algorithm on its classical CartPole benchmark. NFQ was a pioneering approach towards modern Deep Reinforcement Learning (Deep RL) in applying multi-layer neural networks to reinforcement learning for real-world control problems. We explore the algorithm's conceptual simplicity and its transition from online to batch learning, which contributed to its stability. Despite its initial success, NFQ required extensive tuning and was not easily reproducible on real-world control problems. We propose a modernized variant NFQ2.0 and apply it to the CartPole task, concentrating on a real-world system build from standard industrial components, to investigate and improve the learning process's repeatability and robustness. Through ablation studies, we highlight key design decisions and hyperparameters that enhance performance and stability of NFQ2.0 over the original variant. Finally, we demonstrate how our findings can assist practitioners in reproducing and improving results and applying deep reinforcement learning more effectively in industrial contexts.