Home > Research > Publications & Outputs > Skeleton-prompt

Electronic data

  • SKELETON_PROMPT

    Accepted author manuscript, 2.65 MB, PDF document

    Available under license: CC BY: Creative Commons Attribution 4.0 International License

Links

Text available via DOI:

View graph of relations

Skeleton-prompt: A cross-dataset transfer learning approach for skeleton action recognition

Research output: Contribution to Journal/MagazineJournal articlepeer-review

E-pub ahead of print
Article number111885
<mark>Journal publication date</mark>31/01/2026
<mark>Journal</mark>Pattern Recognition
Volume169
Publication StatusE-pub ahead of print
Early online date9/06/25
<mark>Original language</mark>English

Abstract

This paper presents Skeleton-Prompt, a novel tuning method designed to tackle cross-dataset transfer issues in skeleton action recognition models. Given the scarcity of large-scale 3D skeleton datasets and the variability in keypoint structures across datasets, existing methods often rely on training models from scratch, necessitating extensive labeled data and exhibiting high sensitivity to occlusion. Our approach aims to fine-tune pre-trained models to adapt to limited real-world skeleton data. We use 2D skeletons as inputs and leverage a large human motion dataset for 2D to 3D pose estimation to learn generalizable motion features. A lightweight prompt generator produces instance-level prompts, and we employ dynamic queries with cross-attention to refine the semantic information of the input data. Additionally, we introduce a joint-enhanced multi-stream fusion mechanism based on self-attention to improve robustness against incomplete skeletons. Skeleton-Prompt represents a significant advancement in efficient fine-tuning for skeleton action recognition, effectively addressing cross-dataset generalization challenges in a data-efficient and parameter-efficient manner.