🤖 AI Summary
French prescription instructions suffer from ambiguity, nonstandard phrasing, and colloquialism, hindering reliable structured extraction and thereby compromising medication safety and clinical decision support.
Method: We propose a hybrid pipeline integrating a lightweight Named Entity Recognition and Linking (NERL) system with a fine-tuned large language model (LLM), employing confidence-driven dynamic routing: high-confidence samples are processed by NERL, while low-confidence ones are delegated to the LLM for fine-grained parsing. The approach synergistically combines prompt engineering and parameter-efficient fine-tuning, augmented by a result fusion mechanism.
Contribution/Results: Our method achieves 91% structured extraction accuracy under low-latency constraints—substantially outperforming pure prompt-based approaches and matching the performance of conventional rule- or model-based systems. It demonstrates strong clinical deployability and scalability, marking the first application of a confidence-driven LLM–NERL collaborative paradigm to French prescription structuring.
📝 Abstract
Automatically structuring posology instructions is essential for improving medication safety and enabling clinical decision support. In French prescriptions, these instructions are often ambiguous, irregular, or colloquial, limiting the effectiveness of classic ML pipelines. We explore the use of Large Language Models (LLMs) to convert free-text posologies into structured formats, comparing prompt-based methods and fine-tuning against a "pre-LLM" system based on Named Entity Recognition and Linking (NERL). Our results show that while prompting improves performance, only fine-tuned LLMs match the accuracy of the baseline. Through error analysis, we observe complementary strengths: NERL offers structural precision, while LLMs better handle semantic nuances. Based on this, we propose a hybrid pipeline that routes low-confidence cases from NERL (<0.8) to the LLM, selecting outputs based on confidence scores. This strategy achieves 91% structuration accuracy while minimizing latency and compute. Our results show that this hybrid approach improves structuration accuracy while limiting computational cost, offering a scalable solution for real-world clinical use.