Skip to yearly menu bar Skip to main content


Poster

Encryption-Friendly LLM Architecture

Donghwan Rho · Taeseong Kim · Minje Park · Jung Woo Kim · Hyunsik Chae · Ernest Ryu · Jung Hee Cheon

Hall 3 + Hall 2B #494
[ ] [ Project Page ]
Thu 24 Apr midnight PDT — 2:30 a.m. PDT

Abstract: Large language models (LLMs) offer personalized responses based on user interactions, but this use case raises serious privacy concerns. Homomorphic encryption (HE) is a cryptographic protocol supporting arithmetic computations in encrypted states and provides a potential solution for privacy-preserving machine learning (PPML). However, the computational intensity of transformers poses challenges for applying HE to LLMs. In this work, we propose a modified HE-friendly transformer architecture with an emphasis on inference following personalized (private) fine-tuning. Utilizing LoRA fine-tuning and Gaussian kernels, we achieve significant computational speedups---6.94× for fine-tuning and 2.3× for inference---while maintaining performance comparable to plaintext models. Our findings provide a viable proof of concept for offering privacy-preserving LLM services in areas where data protection is crucial. Our code is available on GitHub.

Live content is unavailable. Log in and register to view live content