-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Poor results when infering #9
Comments
Hi @TheoCarme , Thanks for your interest in our work. You are right, the depth image needs to be normalized to the centre of the cropping box. You can find compute the COM like here then normalise the image like here. You may need to tune the cube size to have the best performance, starting with the default value of the trained dataset of the model. Also, be sure to denormalise the output uvd like here before you draw a figure. Hope these help. |
Thank you for your help. Can you please give me examples of value and shape of cube_size ? I searched but did not find where I could get those default values. Is it right to say that in my situation the cube_size should approximately match with the size of the area in which my hands are present. |
cube_size is a scalar that represents half the length of the side of the square. The default value of the models you use should be 150 (mm). In my experience, the best value is slightly larger than the size of the hand and depends on how good the segmentations are. |
I did not fully understand how the code you suggested work. Yet I tried to use it but did not succeeded. Instead I used this function to normalize my cropped image before resizing it to 128*128. Then I made this function to denormalize the uvd values. Do you see any error in what I did ? Thank you for your time. |
I am sorry for your confusion. In the first part, the denormalization code raise error because the box_size is in the wrong shape. box_size suppose to have a shape of [batch_size], which in your case is [1]. Please follow the procedure in the Are these clear to you? |
It is clear, thank you, I will try that. |
So I made this function to crop and normalize my images. So can you help me understand what the problem is, please. Also with a cube_size of 10000 the cropping function get my images from a shape of about 700700 to 2020. Is this standard ? |
Could you share more information about this error, e.g. which line triggered this error? Also, can you check all the shapes of the input tensors? A |
Hello,
I am currently trying to implement a program that use Pixelwise Regression to estimate the pose of at least two hands on a depth video stream (one frame at the time).
I am using the Stereo Labs ZED Mini camera.
Since Pixelwise can only estimate the pose of only one hand on a frame, I begin by using Mediapipe Hands (I know it is overkill, I may change later) to locate the hands and crop them on the frame. Then I resize the cropped hands to a size of 128x128.
Finally I can use this code :
To get this
After I looked into the values of img, label_img and mask when executing test_samples.py, I got the feeling that contrary to mine those matrix are normalized, which could be the cause of my poor results.
So is my feeling right and if so can you explain to me how I can do the same treatment on my matrix.
P.S. : I tested with both HAND17 and MSRA pretrained models.
Thank you for your work.
The text was updated successfully, but these errors were encountered: