Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Semantic Re-tuning with Contrastive Tension
RISE Research Institutes of Sweden, Digital Systems, Data Science.ORCID iD: 0000-0003-2811-7481
RISE Research Institutes of Sweden, Digital Systems, Data Science.ORCID iD: 0000-0002-9162-6433
RISE Research Institutes of Sweden, Digital Systems, Data Science.
RISE Research Institutes of Sweden, Digital Systems, Data Science.
Show others and affiliations
2021 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Extracting semantically useful natural language sentence representations frompre-trained deep neural networks such as Transformers remains a challenge. Wefirst demonstrate that pre-training objectives impose a significant task bias ontothe final layers of models, with a layer-wise survey of the Semantic Textual Similarity (STS) correlations for multiple common Transformer language models. Wethen propose a new self-supervised method called Contrastive Tension (CT) tocounter such biases. CT frames the training objective as a noise-contrastive taskbetween the final layer representations of two independent models, in turn makingthe final layer representations suitable for feature extraction. Results from multiple common unsupervised and supervised STS tasks indicate that CT outperformsprevious State Of The Art (SOTA), and when combining CT with supervised datawe improve upon previous SOTA results with large margins.

Place, publisher, year, edition, pages
2021.
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:ri:diva-59816OAI: oai:DiVA.org:ri-59816DiVA, id: diva2:1684806
Conference
International Conference on Learning Representations, 2021
Available from: 2022-07-28 Created: 2022-07-28 Last updated: 2024-05-15

Open Access in DiVA

fulltext(2274 kB)422 downloads
File information
File name FULLTEXT01.pdfFile size 2274 kBChecksum SHA-512
cf34ca1a6bec3a38c2041f8f04a25ca96c02667c90d1dc5dc4b58d1066fe88d177a9b93ffdf3ab8c8efcb24df1f6f64f2f82bb1377a78a957e8b4f1849ed4fca
Type fulltextMimetype application/pdf

Authority records

Carlsson, FredrikGogoulou, EvangeliaSahlgren, Magnus

Search in DiVA

By author/editor
Carlsson, FredrikGogoulou, EvangeliaSahlgren, Magnus
By organisation
Data Science
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 422 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 1453 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf