Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning in the Dark: Privacy-Preserving Machine Learning using Function Approximation
Tampere University, Finland.
RISE Research Institutes of Sweden. Tampere University, Finland.
2024 (English)In: 2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, IEEE COMPUTER SOC , 2024, p. 62-71Conference paper, Published paper (Refereed)
Abstract [en]

Over the past few years, a tremendous growth of machine learning was brought about by a significant increase in adoption and implementation of cloud-based services. As a result, various solutions have been proposed in which the machine learning models run on a remote cloud provider and not locally on a user’s machine. However, when such a model is deployed on an untrusted cloud provider, it is of vital importance that the users’ privacy is preserved. To this end, we propose Learning in the Dark - a hybrid machine learning model in which the training phase occurs in plaintext data, but the classification of the users’ inputs is performed directly on homomorphically encrypted ciphertexts. To make our construction compatible with homomorphic encryption, we approximate the ReLU and Sigmoid activation functions using low-degree Chebyshev polynomials. This allowed us to build Learning in the Dark - a privacy-preserving machine learning model that can classify encrypted images with high accuracy. Learning in the Dark preserves users’ privacy since it is capable of performing high accuracy predictions by performing computations directly on encrypted data. In addition to that, the output of Learning in the Dark is generated in a blind and therefore privacy-preserving way by utilizing the properties of homomorphic encryption.

Place, publisher, year, edition, pages
IEEE COMPUTER SOC , 2024. p. 62-71
Series
IEEE International Conference on Trust Security and Privacy in Computing and Communications
Keywords [en]
Activation Function; Homomorphic Encryption; Neural Networks; Polynomial Approximation; Privacy
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:ri:diva-77421DOI: 10.1109/TrustCom60117.2023.00031OAI: oai:DiVA.org:ri-77421DiVA, id: diva2:1937014
Conference
IEEE 22nd International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom) / BigDataSE Conference / CSE Conference / EUC Conference / ISCI Conference, Exeter, ENGLAND, NOV 01-03, 2023
Available from: 2025-02-12 Created: 2025-02-12 Last updated: 2025-09-23Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text
By organisation
RISE Research Institutes of Sweden
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 10 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf