Document Type
Conference Paper
Publication Date
2024
DOI
10.1145/3634737.3637680
Publication Title
ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security
Pages
1034-1048
Conference Name
19th ACM Asia Conference on Computer and Communications Security, 1-5 July 2024, Singapore, Singapore
Abstract
To enable common users to capitalize on the power of deep learning, Machine Learning as a Service (MLaaS) has been proposed in the literature, which opens powerful deep learning models of service providers to the public. To protect the data privacy of end users, as well as the model privacy of the server, several state-of-the-art privacy-preserving MLaaS frameworks have also been proposed. Nevertheless, despite the exquisite design of these frameworks to enhance computation efficiency, the computational cost remains expensive for practical applications. To improve the computation efficiency of deep learning (DL) models, model pruning has been adopted as a strategic approach to remarkably compress DL models. However, for practical deep neural networks, a problem called pruning structure inflation significantly limits the pruning efficiency, as it can seriously hurt the model accuracy. In this paper, we propose MOSAIC, a highly flexible pruning framework, to address this critical challenge. By first pruning the network with the carefully selected basic pruning units, then assembling the pruned units into suitable HE Pruning Structures through smart channel transformations, MOSAIC achieves a high pruning ratio while avoiding accuracy reduction, eliminating the problem plagued by the pruning structure inflation. We apply MOSAIC to popular DL models such as VGG and ResNet series on classic datasets such as CIFAR-10 and Tiny ImageNet. Experimental results demonstrate that MOSAIC effectively and flexibly conducts pruning on those models, significantly reducing the Perm, Mult, and Add operations to achieve the global cost reduction without any loss in accuracy. For instance, in VGG-16 on Tiny ImageNet, the total cost is reduced to 21.14% and 29.49% under the MLaaS frameworks GAZELLE and CrypTFlow2, respectively.
Rights
© 2024 Held by the Owner/Authors.
This work is licensed under a Creative Commons Attribution International 4.0 (CC BY 4.0) License.
Original Publication Citation
Cai, Y., Zhang, Q., Ning, R., Xin, C., & Wu, H. (2024) MOSAIC: A prune-and-assemble approach for efficient model pruning in privacy-preserving deep learning. In ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security (pp 1034-1048). Association for Computing Machinery. https://doi.org/10.1145/3634737.3637680
Repository Citation
Cai, Y., Zhang, Q., Ning, R., Xin, C., & Wu, H. (2024) MOSAIC: A prune-and-assemble approach for efficient model pruning in privacy-preserving deep learning. In ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security (pp 1034-1048). Association for Computing Machinery. https://doi.org/10.1145/3634737.3637680
ORCID
0000-0003-4050-6252 (Ning)
Included in
Artificial Intelligence and Robotics Commons, Electrical and Computer Engineering Commons, Information Security Commons