TY - GEN
T1 - Embodied gesture learning from one-shot
AU - Cabrera, Maria E.
AU - Wachs, Juan P.
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2016/11/15
Y1 - 2016/11/15
N2 - This paper discusses the problem of one shot gesture recognition. This is relevant to the field of human-robot interaction, where the user's intentions are indicated through spontaneous gesturing (one shot) to the robot. The novelty of this work consists of learning the process that leads to the creation of a gesture, rather on the gesture itself. In our case, the context involves the way in which humans produce the gestures - the kinematic and anthropometric characteristics and the users' proxemics (the use of the space around them). In the method presented, the strategy is to generate a dataset of realistic samples based on biomechanical features extracted from a single gesture sample. These features, called 'the gist of a gesture', are considered to represent what humans remember when seeing a gesture and the cognitive process involved when trying to replicate it. By adding meaningful variability to these features, a large training data set is created while preserving the fundamental structure of the original gesture. Having a large dataset of realistic samples enables training classifiers for future recognition. Three classifiers were trained and tested using a subset of ChaLearn dataset, resulting in all three classifiers showing rather similar performance around 80% recognition rate Our classification results show the feasibility and adaptability of the presented technique regardless of the classifier.
AB - This paper discusses the problem of one shot gesture recognition. This is relevant to the field of human-robot interaction, where the user's intentions are indicated through spontaneous gesturing (one shot) to the robot. The novelty of this work consists of learning the process that leads to the creation of a gesture, rather on the gesture itself. In our case, the context involves the way in which humans produce the gestures - the kinematic and anthropometric characteristics and the users' proxemics (the use of the space around them). In the method presented, the strategy is to generate a dataset of realistic samples based on biomechanical features extracted from a single gesture sample. These features, called 'the gist of a gesture', are considered to represent what humans remember when seeing a gesture and the cognitive process involved when trying to replicate it. By adding meaningful variability to these features, a large training data set is created while preserving the fundamental structure of the original gesture. Having a large dataset of realistic samples enables training classifiers for future recognition. Three classifiers were trained and tested using a subset of ChaLearn dataset, resulting in all three classifiers showing rather similar performance around 80% recognition rate Our classification results show the feasibility and adaptability of the presented technique regardless of the classifier.
UR - http://www.scopus.com/inward/record.url?scp=85002644879&partnerID=8YFLogxK
U2 - 10.1109/ROMAN.2016.7745244
DO - 10.1109/ROMAN.2016.7745244
M3 - Conference contribution
AN - SCOPUS:85002644879
T3 - 25th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2016
SP - 1092
EP - 1097
BT - 25th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2016
PB - Institute of Electrical and Electronics Engineers
T2 - 25th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2016
Y2 - 26 August 2016 through 31 August 2016
ER -