You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I notice that in Fig.1 of the technique report, the neutral mesh is used to calculated the mesh offset during audio2lmk training. It seems that you use mediapipe to obtain the mesh vertices in this step, would you please share how you obtained neutral mesh vertices in this step?
I also notice that during audio2vid inference, you add the predicted mesh offset to the non-neutral mesh vertices obtained from the reference image, will this effect the performance of lip sync?
The text was updated successfully, but these errors were encountered:
Thanks for open-sourcing such a wonderful work!
I notice that in Fig.1 of the technique report, the neutral mesh is used to calculated the mesh offset during audio2lmk training. It seems that you use mediapipe to obtain the mesh vertices in this step, would you please share how you obtained neutral mesh vertices in this step?
I also notice that during audio2vid inference, you add the predicted mesh offset to the non-neutral mesh vertices obtained from the reference image, will this effect the performance of lip sync?
The text was updated successfully, but these errors were encountered: