facebookresearch / multiface

Hosts the Multiface dataset, which is a multi-view dataset of multiple identities performing a sequence of facial expressions.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

How to Feed Position Map to Network in PiCA

skyir0n opened this issue · comments

Hello, I'm really interested in your work, especially Pixel Codec Avatar (PiCA).
Actually, it is very attractive work for me.

So, I'd like to ask a question regarding to the PiCA.

A coarse mesh is converted to position map and it is fed up into the encoder, instead of the mesh itself.
The position map obtained from the mesh, has high intensities of x, y, and z coordinates, respectively.
Generally, values of input data fed into a neural network belong to specific ranges. For example, RGB image has values ranging from 0 to 1.

Could you let me know how to feed the position map into the network?
Is a preprocessing needed such as normalization?

I would appreciate if you responds.
Thank you.

Hi,

Thanks for your question! The positional map being fed into the network would be the residual of vertices' tracked coordinates relative to the their mean coordinates on the mesh. This naturally confines them into a particular range.

For details, please refer to the original paper on PiCA: https://research.facebook.com/publications/pixel-codec-avatars/