Endre søk
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
The Inhibitor: ReLU and Addition-Based Attention for Efficient Transformers
RISE Research Institutes of Sweden, Digitala system, Datavetenskap. Luleå University of Technology, Sweden.ORCID-id: 0000-0003-4293-6408
2024 (engelsk)Inngår i: Proceedings of the AAAI Conference on Artificial Intelligence, Association for the Advancement of Artificial Intelligence , 2024, Vol. 38, nr 21, s. 23445-23446Konferansepaper, Publicerat paper (Fagfellevurdert)
Abstract [en]

To enhance the computational efficiency of quantized Transformers, we replace the dot-product and Softmax-based attention with an alternative mechanism involving addition and ReLU activation only. This side-steps the expansion to double precision often required by matrix multiplication and avoids costly Softmax evaluations but maintains much of the core functionality of conventional dot-product attention. It can enable more efficient execution and support larger quantized Transformer models on resource-constrained hardware or alternative arithmetic systems like homomorphic encryption. Training experiments on four common benchmark tasks show test set prediction scores comparable to those of conventional Transformers with dot-product attention. Our scaling experiments also suggest significant computational savings, both in plaintext and under encryption. The ReLU and addition-based attention mechanism introduced in this paper may enable privacy-preserving AI applications operating under homomorphic encryption by avoiding the costly multiplication of encrypted variables.

sted, utgiver, år, opplag, sider
Association for the Advancement of Artificial Intelligence , 2024. Vol. 38, nr 21, s. 23445-23446
Emneord [en]
Artificial intelligence; Computational efficiency; Computational savings; Conventional transformer; Core functionality; Double precision; Ho-momorphic encryptions; Homomorphic-encryptions; MAtrix multiplication; Scaling experiments; Test sets; Transformer modeling; Cryptography
HSV kategori
Identifikatorer
URN: urn:nbn:se:ri:diva-72840DOI: 10.1609/aaai.v38i21.30422Scopus ID: 2-s2.0-85189627116OAI: oai:DiVA.org:ri-72840DiVA, id: diva2:1854899
Konferanse
38th AAAI Conference on Artificial Intelligence, AAAI 2024. Vancouver, Canada. 20 February 2024 through 27 February 2024
Tilgjengelig fra: 2024-04-29 Laget: 2024-04-29 Sist oppdatert: 2024-04-29bibliografisk kontrollert

Open Access i DiVA

Fulltekst mangler i DiVA

Andre lenker

Forlagets fulltekstScopus

Person

Brännvall, Rickard

Søk i DiVA

Av forfatter/redaktør
Brännvall, Rickard
Av organisasjonen

Søk utenfor DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric

doi
urn-nbn
Totalt: 201 treff
RefereraExporteraLink to record
Permanent link

Direct link
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annet format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annet språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
v. 2.46.0