Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

resolution ratio of input image #23

Open
chky1997 opened this issue Feb 27, 2023 · 5 comments
Open

resolution ratio of input image #23

chky1997 opened this issue Feb 27, 2023 · 5 comments

Comments

@chky1997
Copy link

chky1997 commented Feb 27, 2023

Hi, it seems a little blurry when using your gui_human.py to visualize the results. Does the resolution ratio (input_ratio in the yaml) that cause the problem? Will the result seems much clearer if the parameter set to 1.0 for training and inference? Thank you!

@haotongl
Copy link
Member

Although I haven't conducted that particular experiment yet, my experience with other datasets suggests that training a model with full views (21 views for ZJU-MoCap) and an input ratio of 1.0 can lead to optimal rendering results.

@chky1997
Copy link
Author

chky1997 commented Mar 3, 2023

About the outdoor dataset, what's the resolution ratio when your cameras record the videos? Do you resize the images to 1024*1024 just after recording, before getting the smpl keypoints?
In project page, the video of outdoor dataset also seems clearer than zjumocap dataset. Is there any difference between the two dataset during the recording stage?

@haotongl
Copy link
Member

haotongl commented Mar 3, 2023

The zjumocap dataset is captured with 21 industrial cameras (2048x2048). We resize the images to 1024*1024.
I think the estimation of smpl keypoints under different resolutions will not affect the rendering results a lot since it is only used to defined a bbox to bound the foreground region.

The outdoor dataset is captured with 18 GoPro Cameras (1920x1080). We keep the original resolution.

@chky1997
Copy link
Author

chky1997 commented Mar 6, 2023

About the outdoor dataset, I found the vhull dir contains the 3D bbox information. But I wonder how to get background.ply. Is it generated from the 18 background images? Also, I noticed outdoor dataset no longer needs the smpl points, it just needs the human images, human 3d mask (generated from 2d mask and converted to 3d using camera intri and extri) and background information, is that right?
By the way, could you tell me the average distance between each gopro cameras, thank you!

@haotongl
Copy link
Member

haotongl commented Mar 6, 2023

  1. Bckground.ply is the SFM sparse point cloud which is computed during calibration.
  2. Outdoor dataset does not needs human mask information. To obtain the 3d bbox, you can follow this suggestion:
    The vedio of the ZJU-Mocap #27 (comment)
  3. About 0.1-0.3m. The specific value can be obtained by calculating the distance between camera positions through extri.yml. Units in Extri.yml have been normalized to meters.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants