Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Component attention network for multimodal dance improvisation recognition
RISE Research Institutes of Sweden, Digital Systems, Data Science. KTH Royal Institute of Technology, Sweden.ORCID iD: 0009-0004-3798-8603
RISE Research Institutes of Sweden.
RISE Research Institutes of Sweden.
RISE Research Institutes of Sweden, Digital Systems, Data Science.ORCID iD: 0000-0003-3272-4145
Show others and affiliations
2023 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Dance improvisation is an active research topic in the arts. Motion analysis of improvised dance can be challenging due to its unique dynamics. Data-driven dance motion analysis, including recognition and generation, is often limited to skeletal data. However, data of other modalities, such as audio, can be recorded and benefit downstream tasks. This paper explores the application and performance of multimodal fusion methods for human motion recognition in the context of dance improvisation. We propose an attention-based model, component attention network (CANet), for multimodal fusion on three levels: 1) feature fusion with CANet, 2) model fusion with CANet and graph convolutional network (GCN), and 3) late fusion with a voting strategy. We conduct thorough experiments to analyze the impact of each modality in different fusion methods and distinguish critical temporal or component features. We show that our proposed model outperforms the two baseline methods, demonstrating its potential for analyzing improvisation in dance

Place, publisher, year, edition, pages
Association for Computing Machinery , 2023. p. 114-118
Keywords [en]
Arts computing; Attention network; Dance recognition; Data driven; Down-stream; Fusion methods; Improvization; Multi-modal; Multi-modal fusion; Performance; Research topics; Motion estimation
National Category
Computer and Information Sciences
Identifiers
URN: urn:nbn:se:ri:diva-67967DOI: 10.1145/3577190.3614114Scopus ID: 2-s2.0-85175844284OAI: oai:DiVA.org:ri-67967DiVA, id: diva2:1814393
Conference
25th International Conference on Multimodal Interaction, ICMI 2023. Paris, France. 9 October 2023 through 13 October 2023
Available from: 2023-11-24 Created: 2023-11-24 Last updated: 2024-02-06Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusFulltext

Authority records

Fu, JiaPashami, Sepideh

Search in DiVA

By author/editor
Fu, JiaPashami, Sepideh
By organisation
Data ScienceRISE Research Institutes of Sweden
Computer and Information Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 30 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf