juanpino2000.bsky.social
@juanpino2000.bsky.social
Current limitations: Our work is still at a research stage and we recognize limitations that still exist: more semantically-relevant gestures (closely linking gesture generation with language modeling (LLM) to allow gestures that are more semantically relevant), low latency, etc.
June 27, 2025 at 6:18 PM
In line with FAIR mission, we open source Seamless Interaction Dataset to the community in order to accelerate the field. Download on github github.com/facebookrese... and @hf.co huggingface.co/datasets/fac...
GitHub - facebookresearch/seamless_interaction: Foundation Models and Data for Human-Human and Human-AI interactions.
Foundation Models and Data for Human-Human and Human-AI interactions. - facebookresearch/seamless_interaction
github.com
June 27, 2025 at 6:16 PM
We’ve trained all our dyadic motion models on Seamless Interaction Dataset. Seamless Interaction Dataset is the first-of-its-kind dataset in terms of scale and breadth, with 4000+ hours and 4000+ participants.
June 27, 2025 at 6:16 PM
The output representations of our motion models are compatible with 2D and 3D renderings.
June 27, 2025 at 6:15 PM
Illustrative Gestures: our models generate gestures that match the semantic content (e.g. extend the arms when saying the word “fly”).
June 27, 2025 at 6:14 PM
Models can be controlled to enable more face expressiveness, with potential applications towards building more attentive or empathetic virtual listeners.
June 27, 2025 at 6:14 PM
The AV Dyadic Motion models go one step further to show visual synchrony by also taking into consideration the visual input of the other party.
June 27, 2025 at 6:13 PM
We built a family of Audio-Visual (AV) Dyadic Motion research models. Our models, conditioned on speech from two parties, can jointly generate facial expressions and body gestures.
June 27, 2025 at 6:12 PM