Breast-Rehab: A Postoperative Breast Cancer Rehabilitation Training Assessment System Based on Human Action Recognition

📅 2025-12-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Low adherence to home-based upper-limb rehabilitation among post-mastectomy breast cancer patients, coupled with the high hardware cost and poor clinical deployability of existing virtual reality solutions, poses a significant barrier to effective recovery. To address this, we propose a low-cost, lightweight home rehabilitation training and assessment system. Our method introduces a novel action segmentation algorithm that fuses visual and 3D skeletal data, and a domain-knowledge-enhanced Retrieval-Augmented Generation (RAG) framework that drives a multimodal large language model to generate clinically credible evaluation reports—substantially reducing computational requirements and hallucination risks. The system integrates a custom lightweight action recognition model, a WeChat Mini-Program for patients, and a nurse-facing management dashboard. Preliminary clinical validation demonstrates an average exercise frequency of 0.59 sessions per day over two weeks, confirming high feasibility and user acceptability. This work provides a scalable, resource-efficient technical pathway for precision rehabilitation in resource-constrained settings.

Technology Category

Application Category

📝 Abstract
Postoperative upper limb dysfunction is prevalent among breast cancer survivors, yet their adherence to at-home rehabilitation exercises is low amidst limited nursing resources. The hardware overhead of commonly adopted VR-based mHealth solutions further hinders their widespread clinical application. Therefore, we developed Breast-Rehab, a novel, low-cost mHealth system to provide patients with out-of-hospital upper limb rehabilitation management. Breast-Rehab integrates a bespoke human action recognition algorithm with a retrieval-augmented generation (RAG) framework. By fusing visual and 3D skeletal data, our model accurately segments exercise videos recorded in uncontrolled home environments, outperforming standard models. These segmented clips, combined with a domain-specific knowledge base, guide a multi-modal large language model to generate clinically relevant assessment reports. This approach significantly reduces computational overhead and mitigates model hallucinations. We implemented the system as a WeChat Mini Program and a nurse-facing dashboard. A preliminary clinical study validated the system's feasibility and user acceptance, with patients achieving an average exercise frequency of 0.59 sessions/day over a two-week period. This work thus presents a complete, validated pipeline for AI-driven, at-home rehabilitation monitoring.
Problem

Research questions and friction points this paper is trying to address.

Addresses low adherence to at-home rehabilitation exercises for breast cancer survivors
Reduces hardware overhead and computational cost of mHealth solutions
Provides accurate AI-driven monitoring and assessment in uncontrolled home environments
Innovation

Methods, ideas, or system contributions that make the work stand out.

Fuses visual and 3D skeletal data for action recognition
Combines segmented clips with knowledge base via RAG framework
Implements system as low-cost WeChat Mini Program and dashboard
🔎 Similar Papers
No similar papers found.
Z
Zikang Chen
Zhejiang University, Hangzhou 310058, China
T
Tan Xie
Zhejiang University School of Medicine Sir Run Run Shaw Hospital, Hangzhou 310058, China
Q
Qinchuan Wang
Zhejiang University School of Medicine Sir Run Run Shaw Hospital, Hangzhou 310058, China
H
Heming Zheng
Zhejiang University School of Medicine Sir Run Run Shaw Hospital, Hangzhou 310058, China
Xudong Lu
Xudong Lu
PhD student, the Chinese University of Hong Kong
Computer VisionMachine Learning