@juanpino2000.bsky.social
18 followers 3 following 10 posts
Posts Media Videos Starter Packs
juanpino2000.bsky.social
Current limitations: Our work is still at a research stage and we recognize limitations that still exist: more semantically-relevant gestures (closely linking gesture generation with language modeling (LLM) to allow gestures that are more semantically relevant), low latency, etc.
juanpino2000.bsky.social
We’ve trained all our dyadic motion models on Seamless Interaction Dataset. Seamless Interaction Dataset is the first-of-its-kind dataset in terms of scale and breadth, with 4000+ hours and 4000+ participants.
juanpino2000.bsky.social
The output representations of our motion models are compatible with 2D and 3D renderings.
juanpino2000.bsky.social
Illustrative Gestures: our models generate gestures that match the semantic content (e.g. extend the arms when saying the word “fly”).
juanpino2000.bsky.social
Models can be controlled to enable more face expressiveness, with potential applications towards building more attentive or empathetic virtual listeners.
juanpino2000.bsky.social
The AV Dyadic Motion models go one step further to show visual synchrony by also taking into consideration the visual input of the other party.
juanpino2000.bsky.social
We built a family of Audio-Visual (AV) Dyadic Motion research models. Our models, conditioned on speech from two parties, can jointly generate facial expressions and body gestures.