Altmetric
Acceleration and design of homomorphically encrypted convolutional neural networks
File | Description | Size | Format | |
---|---|---|---|---|
Chua-Z-2024-PhD-Thesis.pdf | Thesis | 10.3 MB | Adobe PDF | View/Open |
Title: | Acceleration and design of homomorphically encrypted convolutional neural networks |
Authors: | Chua, Zhi Ming |
Item Type: | Thesis or dissertation |
Abstract: | In recent years, there has been a steady increase in the utilisation of third-party servers equipped with large computational resource for computation outsourcing. This has raised serious concerns regarding data security and privacy. A potential solution is homomorphic encryption, which allows operations to be performed directly on encrypted data, thereby ensuring that the underlying data is never exposed during computation. However, the practicality of homomorphic encryption remains restricted because homomorphic operations are significantly more computationally expensive than their plaintext counterparts. This is especially challenging for machine learning applications, which generally involve computation-intensive workloads. In this thesis, we address the problem of accelerating the homomorphic computation of convolutional neural networks and explore the design of low-latency neural networks evaluated under homomorphic encryption. We begin by focusing on accelerating the homomorphic computation of individual convolutional layers by leveraging batching, which is a capability of certain homomorphic encryption schemes. Batching allows multiple data values to be packed into the same ciphertext for concurrent computation, reducing the number of homomorphic operations required. Additionally, we propose partitioning homomorphically encrypted convolutional neural networks into sub-networks. We consider the search for an optimal partition that balances the communication and computation costs to yield the lowest overall latency. We also introduce a framework to systematically traverse the design space, exploring different ways to partition a network and pack data into ciphertexts, in search for a design that minimises the overall homomorphic latency. To facilitate quick homomorphic latency estimation in the neural architecture search, we develop a tool that extrapolates the latency of small neural networks to predict the homomorphic latency of a given neural network. Our findings demonstrate that applying neural architecture search to homomorphically encrypted neural networks in an homomorphic encryption-aware manner yields significant performance gains when compared to using plaintext proxies in the performance estimation step. |
Content Version: | Open Access |
Issue Date: | Dec-2023 |
Date Awarded: | Mar-2024 |
URI: | http://hdl.handle.net/10044/1/110324 |
DOI: | https://doi.org/10.25560/110324 |
Copyright Statement: | Creative Commons Attribution NonCommercial Licence |
Supervisor: | Cheung, Peter Bouganis, Christos-Savvas |
Department: | Department of Electrical and Electronic Engineering |
Publisher: | Imperial College London |
Qualification Level: | Doctoral |
Qualification Name: | Doctor of Philosophy (PhD) |
Appears in Collections: | Electrical and Electronic Engineering PhD theses |
This item is licensed under a Creative Commons License