Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How can I know which face on the video sample is speaking to extract my own visual features ?? #50

Open
david-gimeno opened this issue Nov 30, 2023 · 1 comment

Comments

@david-gimeno
Copy link

First of all, I would like to congratulate all you for the big effort you did when creating this MELD dataset. However, I would also like to ask you if it is possible to obtain the facial landmarks (or any other kind of information) that will allow me to extract the face of the person actively speaking as you did for extracting the features you provide.

The reason is because I would like to explore my own visual features.

Thanks in advance. Best regards from Valencia,

David

@rajendrac3
Copy link

Hi,

I am also trying to get the faces which are speaking in the video.
This research paper does something similar https://arxiv.org/pdf/2101.03149
Here is the code implementation: https://github.com/facebookresearch/VisualVoice/tree/main

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants