Confidential Prompting: Protecting User Prompts from Cloud LLM Providers

📅 2024-09-27
🏛️ arXiv.org
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
To address the risk of sensitive user prompts—such as clinical records or financial data—being leaked in cloud-hosted large language model (LLM) services, this paper proposes an end-to-end privacy-preserving framework balancing prompt confidentiality, output consistency, and computational efficiency. The method integrates secure prompt sharding and decoding (SPD) with prompt obfuscation (PO) within a trusted execution environment (TEE), leveraging token-level secure decoupling to achieve absolute prompt isolation inside confidential VMs and robustness against prompt reconstruction attacks. Evaluation demonstrates that the framework preserves original LLM output quality and inference latency while providing strong confidentiality guarantees for sensitive prompts. It significantly outperforms existing approaches based on homomorphic encryption or differential privacy in both security and efficiency.

Technology Category

Application Category

📝 Abstract
Our work tackles the challenge of securing user inputs in cloud-hosted large language model (LLM) serving while ensuring model confidentiality, output invariance, and compute efficiency. We introduce Secure Partitioned Decoding (SPD), which uses confidential computing to confine user prompts to a trusted execution environment (TEE), namely a confidential virtual machine (CVM), while allowing service providers to generate tokens efficiently. We also introduce a novel cryptographic method, Prompt Obfuscation (PO), to ensure robustness against reconstruction attacks on SPD. We demonstrate our approach preserves both prompt confidentiality and LLM serving efficiency. Our solution enables privacy-preserving cloud LLM serving that handles sensitive prompts, such as clinical records, financial data, and personal information.
Problem

Research questions and friction points this paper is trying to address.

Securing user inputs in cloud-hosted LLM serving
Ensuring model confidentiality and output invariance
Protecting sensitive prompts like clinical and financial data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Secure Partitioned Decoding using confidential computing
Prompt Obfuscation for robustness against attacks
Confidential virtual machine for secure prompt handling
🔎 Similar Papers
No similar papers found.
I
In Gim
Department of Computer Science, Yale University
C
Caihua Li
Department of Computer Science, Yale University
L
Lin Zhong
Department of Computer Science, Yale University