STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video
Abstract In recent years, video‐based hand–object interaction has received widespread attention from researchers. However, due to the complexity and occlusion of hand movements, hand–object interaction recognition based on RGB videos remains a highly challenging task. Here, an end‐to‐end spatio‐temp...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2024-09-01
|
Series: | Electronics Letters |
Subjects: | |
Online Access: | https://doi.org/10.1049/ell2.70010 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1846173408196296704 |
---|---|
author | Jiao Liang Xihan Wang Jiayi Yang Quanli Gao |
author_facet | Jiao Liang Xihan Wang Jiayi Yang Quanli Gao |
author_sort | Jiao Liang |
collection | DOAJ |
description | Abstract In recent years, video‐based hand–object interaction has received widespread attention from researchers. However, due to the complexity and occlusion of hand movements, hand–object interaction recognition based on RGB videos remains a highly challenging task. Here, an end‐to‐end spatio‐temporal former (STFormer) network for understanding hand behaviour in interactions is proposed. The network consists of three modules: FlexiViT feature extraction, hand–object pose estimator, and interaction action classifier. The FlexiViT is used to extract multi‐scale features from each image frame. The hand–object pose estimator is designed to predict 3D hand pose keypoints and object labels for each frame. The interaction action classifier is used to predict the interaction action categories for the entire video. The experimental results demonstrate that our approach achieves competitive recognition accuracies of 94.96% and 88.84% on two datasets, namely first‐person hand action (FPHA) and 2 Hands and Objects (H2O). |
format | Article |
id | doaj-art-77d733feffb74d2bb88a4d79f2ef2aa4 |
institution | Kabale University |
issn | 0013-5194 1350-911X |
language | English |
publishDate | 2024-09-01 |
publisher | Wiley |
record_format | Article |
series | Electronics Letters |
spelling | doaj-art-77d733feffb74d2bb88a4d79f2ef2aa42024-11-08T14:35:49ZengWileyElectronics Letters0013-51941350-911X2024-09-016017n/an/a10.1049/ell2.70010STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB videoJiao Liang0Xihan Wang1Jiayi Yang2Quanli Gao3State‐Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services Xi'an Polytechnic University Xi'an ChinaState‐Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services Xi'an Polytechnic University Xi'an ChinaState‐Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services Xi'an Polytechnic University Xi'an ChinaState‐Province Joint Engineering and Research Center of Advanced Networking and Intelligent Information Services Xi'an Polytechnic University Xi'an ChinaAbstract In recent years, video‐based hand–object interaction has received widespread attention from researchers. However, due to the complexity and occlusion of hand movements, hand–object interaction recognition based on RGB videos remains a highly challenging task. Here, an end‐to‐end spatio‐temporal former (STFormer) network for understanding hand behaviour in interactions is proposed. The network consists of three modules: FlexiViT feature extraction, hand–object pose estimator, and interaction action classifier. The FlexiViT is used to extract multi‐scale features from each image frame. The hand–object pose estimator is designed to predict 3D hand pose keypoints and object labels for each frame. The interaction action classifier is used to predict the interaction action categories for the entire video. The experimental results demonstrate that our approach achieves competitive recognition accuracies of 94.96% and 88.84% on two datasets, namely first‐person hand action (FPHA) and 2 Hands and Objects (H2O).https://doi.org/10.1049/ell2.70010computer visionimage classificationpose estimation |
spellingShingle | Jiao Liang Xihan Wang Jiayi Yang Quanli Gao STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video Electronics Letters computer vision image classification pose estimation |
title | STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video |
title_full | STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video |
title_fullStr | STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video |
title_full_unstemmed | STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video |
title_short | STFormer: Spatio‐temporal former for hand–object interaction recognition from egocentric RGB video |
title_sort | stformer spatio temporal former for hand object interaction recognition from egocentric rgb video |
topic | computer vision image classification pose estimation |
url | https://doi.org/10.1049/ell2.70010 |
work_keys_str_mv | AT jiaoliang stformerspatiotemporalformerforhandobjectinteractionrecognitionfromegocentricrgbvideo AT xihanwang stformerspatiotemporalformerforhandobjectinteractionrecognitionfromegocentricrgbvideo AT jiayiyang stformerspatiotemporalformerforhandobjectinteractionrecognitionfromegocentricrgbvideo AT quanligao stformerspatiotemporalformerforhandobjectinteractionrecognitionfromegocentricrgbvideo |