Scalable23

From ACES

Revision as of 08:23, 10 July 2023 by X5zhang (talk | contribs) (Created page with "{{publication |entry=article |year=2023 |month=7 |author=Xinqiao Zhang, Mohammad Samragh, Siam Hussain, Ke Huang, Farinaz Koushanfar |title=Scalable Binary Neural Network appl...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Scalable23
entryarticle
address
annote
authorXinqiao Zhang, Mohammad Samragh, Siam Hussain, Ke Huang, Farinaz Koushanfar
booktitle
chapter
edition
editor
howpublished
institution
journal
month7
note
number
organization
pages
publisherACM Transactions on Embedded Computing Systems
school
series
titleScalable Binary Neural Network applications in Oblivious Inference
type
volume
year2023
doi
issn
isbn
urlhttps://dl.acm.org/doi/pdf/10.1145/3607192
pdfScalable23

page=1

Binary neural network (BNN) delivers increased compute intensity and reduces memory/data requirements for computation. Scalable BNN enables inference in a limited time due to diferent constraints. This paper explores the application of Scalable BNN in oblivious inference, a service provided by a server to mistrusting clients. Using this service, a client can obtain the inference result on his/her data by a trained model held by the server without disclosing the data or learning the model parameters. Two contributions of this paper are: 1) we devise lightweight cryptographic protocols explicitly designed to exploit the unique characteristics of BNNs. 2) we present an advanced dynamic exploration of the runtime-accuracy tradeof of scalable BNNs in a single-shot training process. While previous works trained multiple BNNs with diferent computational complexities (which is cumbersome due to the slow convergence of BNNs), we train a single BNN that can perform inference under various computational budgets. Compared to CryptFlow2, the state-of-the-art technique in the oblivious inference of non-binary DNNs, our approach reaches 3× faster inference while keeping the same accuracy. Compared to XONN, the state-of-the-art technique in the oblivious inference of binary networks, we achieve 2× to 12× faster inference while obtaining higher accuracy.

Icon-email.png
Email:
farinaz@ucsd.edu
Icon-addr.png
Address:
Electrical & Computer Engineering
University of California, San Diego
9500 Gilman Drive, MC 0407
Jacobs Hall, Room 6401
La Jolla, CA 92093-0407
Icon-addr.png
Lab Location: EBU1-2514
University of California San Diego
9500 Gilman Dr, La Jolla, CA 92093