Abstract

Gesture and multimodal communication researchers typically annotate video data manually, even though this can be a very time-consuming task. In the present work, a method to detect gestures is proposed as a fundamental step towards a semi-automatic gesture annotation tool. The proposed method can be applied to RGB videos and requires annotations of part of a video as input. The technique deploys a pose estimation method and active learning. In the experiment, it is shown that if about 27% of the video is annotated, the remaining parts of the video can be annotated automatically with an F-score of at least 0.85. Users can run this tool with a small number of annotations first. If the predicted annotations for the remainder of the video are not satisfactory, users can add further annotations and run the tool again. The code has been released so that other researchers and practitioners can use the results of this research. This tool has been confirmed to work in conjunction with ELAN.

Details

Title
Semi-automation of gesture annotation by machine learning and human collaboration
Author
Ienaga, Naoto 1   VIAFID ORCID Logo  ; Cravotta, Alice 2 ; Terayama, Kei 3   VIAFID ORCID Logo  ; Scotney, Bryan W. 4   VIAFID ORCID Logo  ; Saito, Hideo 1   VIAFID ORCID Logo  ; Busà, M. Grazia 2   VIAFID ORCID Logo 

 Keio University, Yokohama, Japan (GRID:grid.26091.3c) (ISNI:0000 0004 1936 9959) 
 University of Padova, Padova, Italy (GRID:grid.5608.b) (ISNI:0000 0004 1757 3470) 
 Yokohama City University, Yokohama, Japan (GRID:grid.268441.d) (ISNI:0000 0001 1033 6139) 
 Ulster University, Jordanstown, Northern Ireland (GRID:grid.12641.30) (ISNI:0000000105519715) 
Pages
673-700
Publication year
2022
Publication date
Sep 2022
Publisher
Springer Nature B.V.
ISSN
1574020X
e-ISSN
1574-0218
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2703667395
Copyright
© The Author(s) 2022. corrected publication 2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.