Scalable23
From ACES
| Scalable23 | |
|---|---|
| entry | article |
| address | |
| annote | |
| author | Xinqiao Zhang and Mohammad Samragh and Siam Hussain and Ke Huang and Farinaz Koushanfar |
| booktitle | |
| chapter | |
| edition | |
| editor | |
| howpublished | |
| institution | |
| journal | |
| month | 7 |
| note | |
| number | |
| organization | |
| pages | |
| publisher | ACM Transactions on Embedded Computing Systems |
| school | |
| series | |
| title | Scalable Binary Neural Network applications in Oblivious Inference |
| type | |
| volume | |
| year | 2023 |
| doi | |
| issn | |
| isbn | |
| url | https://dl.acm.org/doi/pdf/10.1145/3607192 |
| Scalable23 | |
Binary neural network (BNN) delivers increased compute intensity and reduces memory/data requirements for computation. Scalable BNN enables inference in a limited time due to diferent constraints. This paper explores the application of Scalable BNN in oblivious inference, a service provided by a server to mistrusting clients. Using this service, a client can obtain the inference result on his/her data by a trained model held by the server without disclosing the data or learning the model parameters. Two contributions of this paper are: 1) we devise lightweight cryptographic protocols explicitly designed to exploit the unique characteristics of BNNs. 2) we present an advanced dynamic exploration of the runtime-accuracy tradeof of scalable BNNs in a single-shot training process. While previous works trained multiple BNNs with diferent computational complexities (which is cumbersome due to the slow convergence of BNNs), we train a single BNN that can perform inference under various computational budgets. Compared to CryptFlow2, the state-of-the-art technique in the oblivious inference of non-binary DNNs, our approach reaches 3× faster inference while keeping the same accuracy. Compared to XONN, the state-of-the-art technique in the oblivious inference of binary networks, we achieve 2× to 12× faster inference while obtaining higher accuracy.