🤖 AI Summary
Existing gait control models predominantly rely on linear, single-timescale assumptions and are extensively validated in controlled laboratory settings, yet exhibit limited generalizability to real-world, complex environments. Method: We propose a nonlinear, multi-timescale foot-placement prediction model that jointly encodes biomechanical states and eye-movement signals. The architecture integrates GRUs and Transformers to capture flexible temporal dependencies across heterogeneous input modalities. Contribution/Results: Our model is the first to empirically reveal that gait control exhibits context-dependent (e.g., terrain complexity) and modality-dependent (e.g., walking vs. running; treadmill vs. overground) temporal structure: complex terrains engage faster-timescale predictions, and visual information precedes body-state cues in predictive control. Experiments demonstrate significant performance gains over conventional linear models across diverse locomotor contexts and sensor modalities, strong cross-context generalization, and—critically—the first quantitative characterization of the central nervous system’s temporal hierarchy in locomotor control.
📝 Abstract
Modeling movement in real-world tasks is a fundamental scientific goal. However, it is unclear whether existing models and their assumptions, overwhelmingly tested in laboratory-constrained settings, generalize to the real world. For example, data-driven models of foot placement control -- a crucial action for stable locomotion -- assume linear and single timescale mappings. We develop nonlinear foot placement prediction models, finding that neural network architectures with flexible input history-dependence like GRU and Transformer perform best across multiple contexts (walking and running, treadmill and overground, varying terrains) and input modalities (multiple body states, gaze), outperforming traditional models. These models reveal context- and modality-dependent timescales: there is more reliance on fast-timescale predictions in complex terrain, gaze predictions precede body state predictions, and full-body state predictions precede center-of-mass-relevant predictions. Thus, nonlinear action prediction models provide quantifiable insights into real-world motor control and can be extended to other actions, contexts, and populations.