Date of Award
Summer 2025
Document Type
Dissertation
Degree Name
Doctor of Philosophy (PhD)
Department
Electrical & Computer Engineering
Program/Concentration
Electrical and Computer Engineering
Committee Director
Chunsheng Xin
Committee Member
Hongyi Wu
Committee Member
Jinag Li
Committee Member
Lusi Li
Abstract
Deep learning (DL) has become a powerful tool for solving complex problems, but developing DL models typically requires vast datasets, high computational resources, and expert knowledge—barriers that limit accessibility. Machine Learning as a Service (MLaaS) addresses this challenge by allowing resource-rich providers to deliver pre-trained DL models as services. However, privacy concerns arise: clients hesitate to share sensitive data, while providers protect their proprietary models. To address this, privacy-preserving MLaaS integrates cryptographic techniques into DL computations, as seen in frameworks like Cryptonets, SecureML, GAZELLE, CrypTFlow2, Cheetah, and BOLT. Among them, Homomorphic Encryption (HE) enables computation on encrypted data but remains computationally expensive. This dissertation presents four novel contributions to enhance HE-based deep learning efficiency:
First, Hunter introduces structured pruning using HE-friendly features—internal/external structure and weight diagonal—to reduce HE operations like Perm, Mult, and Add while maintaining accuracy. For instance, on VGG-16 (ImageNet), Hunter cuts Perm to 2% and Mult/Add to 14%. Second, MOSAIC addresses pruning structure inflation by assembling compact, HE-friendly units via intelligent channel transformation, reducing costs by 21.14% and 29.49% on VGG-16 under GAZELLE and CrypTFlow2, respectively. StriaNet designs HE-efficient networks from scratch using StriaBlock, which includes ExPerm-Free Convolution and Cross Kernel modules. It minimizes Perm operations and adapts dimensions to HE characteristics, achieving up to 9.78× speedup on ImageNet and 9.24× on CIFAR-10 over ResNet. Lastly, SecDTD enables efficient Transformer inference via early-stage dynamic token dropping. By integrating Pre-Softmax Scoring (MCN) and Fast Oblivious Median Selection (OMSel), it reduces overhead and achieves up to 16.9× speedup. On BERT-base, it delivers up to 4.47× acceleration on GLUE datasets under BOLT and BumbleBee.
Together, these contributions significantly advance the efficiency and practicality of privacy-preserving deep learning, providing a robust foundation for scalable and efficient privacy-preserving MLaaS frameworks.
Rights
In Copyright. URI: http://rightsstatements.org/vocab/InC/1.0/ This Item is protected by copyright and/or related rights. You are free to use this Item in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s).
DOI
10.25777/41a9-j271
ISBN
9798293842315
Recommended Citation
Cai, Yifei.
"Towards Efficient Privacy-Preserving Deep Learning: HE-Friendly Structures, Flexible Pruning, HE-Efficient Architectures, and Secure Transformer Token Drop"
(2025). Doctor of Philosophy (PhD), Dissertation, Electrical & Computer Engineering, Old Dominion University, DOI: 10.25777/41a9-j271
https://digitalcommons.odu.edu/ece_etds/609
ORCID
0000-0002-1372-656X