Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers
RISE Research Institutes of Sweden, Digitala system, Datavetenskap. Luleå University of Technology, Sweden.ORCID-id: 0000-0003-4293-6408
2024 (Engelska)Ingår i: Proceedings of the AAAI Conference on Artificial Intelligence, Association for the Advancement of Artificial Intelligence , 2024, Vol. 38, nr 21, s. 23445-23446Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

To enhance the computational efficiency of quantized Transformers, we replace the dot-product and Softmax-based attention with an alternative mechanism involving addition and ReLU activation only. This side-steps the expansion to double precision often required by matrix multiplication and avoids costly Softmax evaluations but maintains much of the core functionality of conventional dot-product attention. It can enable more efficient execution and support larger quantized Transformer models on resource-constrained hardware or alternative arithmetic systems like homomorphic encryption. Training experiments on four common benchmark tasks show test set prediction scores comparable to those of conventional Transformers with dot-product attention. Our scaling experiments also suggest significant computational savings, both in plaintext and under encryption. The ReLU and addition-based attention mechanism introduced in this paper may enable privacy-preserving AI applications operating under homomorphic encryption by avoiding the costly multiplication of encrypted variables.

Ort, förlag, år, upplaga, sidor
Association for the Advancement of Artificial Intelligence , 2024. Vol. 38, nr 21, s. 23445-23446
Nyckelord [en]
Artificial intelligence; Computational efficiency; Computational savings; Conventional transformer; Core functionality; Double precision; Ho-momorphic encryptions; Homomorphic-encryptions; MAtrix multiplication; Scaling experiments; Test sets; Transformer modeling; Cryptography
Nationell ämneskategori
Data- och informationsvetenskap
Identifikatorer
URN: urn:nbn:se:ri:diva-72840DOI: 10.1609/aaai.v38i21.30422Scopus ID: 2-s2.0-85189627116OAI: oai:DiVA.org:ri-72840DiVA, id: diva2:1854899
Konferens
38th AAAI Conference on Artificial Intelligence, AAAI 2024. Vancouver, Canada. 20 February 2024 through 27 February 2024
Tillgänglig från: 2024-04-29 Skapad: 2024-04-29 Senast uppdaterad: 2024-04-29Bibliografiskt granskad

Open Access i DiVA

Fulltext saknas i DiVA

Övriga länkar

Förlagets fulltextScopus

Person

Brännvall, Rickard

Sök vidare i DiVA

Av författaren/redaktören
Brännvall, Rickard
Av organisationen
Datavetenskap
Data- och informationsvetenskap

Sök vidare utanför DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetricpoäng

doi
urn-nbn
Totalt: 199 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf