Learning Individual Styles of Conversational Gesture
Ground truth
Code (set theory)
DOI:
10.48550/arxiv.1906.04160
Publication Date:
2019-01-01
AUTHORS (6)
ABSTRACT
Human speech is often accompanied by hand and arm gestures. Given audio input, we generate plausible gestures to go along with the sound. Specifically, perform cross-modal translation from "in-the-wild'' monologue of a single speaker their motion. We train on unlabeled videos for which only have noisy pseudo ground truth an automatic pose detection system. Our proposed model significantly outperforms baseline methods in quantitative comparison. To support research toward obtaining computational understanding relationship between gesture speech, release large video dataset person-specific The project website video, code data can be found at http://people.eecs.berkeley.edu/~shiry/speech2gesture .
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....