🤖 AI Summary
Large language models (LLMs) in digital forensics face critical challenges—including bias, poor interpretability, weak admissibility in court, high hardware overhead, and ethical–legal risks. Method: This work systematically defines, for the first time, the applicability boundaries and core risks of LLMs across forensic stages (evidence identification, log parsing, report generation) and proposes a judicially grounded evaluation framework and trust-enhancement principles. It integrates prompt engineering, retrieval-augmented generation (RAG), domain-specific knowledge injection, and explainability analysis to ensure chain-of-custody integrity and evidentiary admissibility. Contribution/Results: Experiments demonstrate a 40% reduction in preliminary analysis time and >92% recall of critical forensic leads. The approach provides both theoretical foundations and a deployable technical paradigm for designing automated, court-compliant forensic tools.