9
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Shredder: Learning Noise to Protect Privacy with Partial DNN Inference on the Edge

      Preprint

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          A wide variety of DNN applications increasingly rely on the cloud to perform their huge computation. This heavy trend toward cloud-hosted inference services raises serious privacy concerns. This model requires the sending of private and privileged data over the network to remote servers, exposing it to the service provider. Even if the provider is trusted, the data can still be vulnerable over communication channels or via side-channel attacks [1,2] at the provider. To that end, this paper aims to reduce the information content of the communicated data without compromising the cloud service's ability to provide a DNN inference with acceptably high accuracy. This paper presents an end-to-end framework, called Shredder, that, without altering the topology or the weights of a pre-trained network, learns an additive noise distribution that significantly reduces the information content of communicated data while maintaining the inference accuracy. Shredder learns the additive noise by casting it as a tensor of trainable parameters enabling us to devise a loss functions that strikes a balance between accuracy and information degradation. The loss function exposes a knob for a disciplined and controlled asymmetric trade-off between privacy and accuracy. While keeping the DNN intact, Shredder enables inference on noisy data without the need to update the model or the cloud. Experimentation with real-world DNNs shows that Shredder reduces the mutual information between the input and the communicated data to the cloud by 70.2% compared to the original execution while only sacrificing 1.46% loss in accuracy.

          Related collections

          Most cited references20

          • Record: found
          • Abstract: not found
          • Article: not found

          Mutual Information and Minimum Mean-Square Error in Gaussian Channels

            Bookmark
            • Record: found
            • Abstract: not found
            • Book Chapter: not found

            Our Data, Ourselves: Privacy Via Distributed Noise Generation

              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              SecureML: A System for Scalable Privacy-Preserving Machine Learning

                Bookmark

                Author and article information

                Journal
                26 May 2019
                Article
                1905.11814
                b086727c-102d-41c8-986c-16cd5c0a2b39

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                cs.CR cs.LG stat.ML

                Security & Cryptology,Machine learning,Artificial intelligence
                Security & Cryptology, Machine learning, Artificial intelligence

                Comments

                Comment on this article