Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you provide the processed data or the related processing code? #7

Open
Haoqing-Wang opened this issue Jul 10, 2023 · 9 comments
Open

Comments

@Haoqing-Wang
Copy link

Nice job! Can you released the used dataset?

@HUAFOR
Copy link

HUAFOR commented Jul 11, 2023

+1

@DavidKong96
Copy link

hope to know which face det model used..

@lmpeng12
Copy link

+1

1 similar comment
@MengShen0709
Copy link

+1

@MengShen0709
Copy link

@DavidKong96
I guess authors might use dlib to extract the landmarks

The partial landmarks are defined in their dataloader:

        landmarks_img = landmarks[13:48]
        landmarks_img2 = landmarks[0:4]
        landmarks_img = np.concatenate((landmarks_img2, landmarks_img))

I use dlib to extracts 68 points, and follow the authors to only keep some of them. It turns out, by doing so, we could remove bottom half of landmarks. This landmark masking is in accordance with the image masking for the model to learn how to generate lip movement.

Here is an illustration:
landmark

But I haven't reproduced the results, so it is just my guess.

Feel free to discuss.

@Haoqing-Wang
Copy link
Author

before obtain landmark, we need to detect the facial RoI advance. But when the model can not detete the face, how to obtain the landmark? we use dlib to obtain landmark.

detector = dlib.get_frontal_face_detector()
predictor = dlib.shape_predictor("shape_predictor_68_face_landmarks.dat")
gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
rect = detector(gray)[0]
shape = predictor(gray, rect)

rect could be None.

@jianmanLin
Copy link

@DavidKong96 I guess authors might use dlib to extract the landmarks

The partial landmarks are defined in their dataloader:

        landmarks_img = landmarks[13:48]
        landmarks_img2 = landmarks[0:4]
        landmarks_img = np.concatenate((landmarks_img2, landmarks_img))

I use dlib to extracts 68 points, and follow the authors to only keep some of them. It turns out, by doing so, we could remove bottom half of landmarks. This landmark masking is in accordance with the image masking for the model to learn how to generate lip movement.

Here is an illustration: landmark

But I haven't reproduced the results, so it is just my guess.

Feel free to discuss.

Hello, did you successfully reproduce this paper? As a result of my training, the inpainting area will keep shaking, and then the training loss will drop rapidly at the beginning, and then it will shake within a small area. I feel very happy troubled

@jianmanLin
Copy link

before obtain landmark, we need to detect the facial RoI advance. But when the model can not detete the face, how to obtain the landmark? we use dlib to obtain landmark.

detector = dlib.get_frontal_face_detector()
predictor = dlib.shape_predictor("shape_predictor_68_face_landmarks.dat")
gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY)
rect = detector(gray)[0]
shape = predictor(gray, rect)

rect could be None.

Hello, did you successfully reproduce this paper? As a result of my training, the inpainting area will keep shaking, and then the training loss will drop rapidly at the beginning, and then it will shake within a small area. I feel very happy troubled

@Utkarsh-shift
Copy link

same request

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants