Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to prepear my dataset? #15

Open
ccsvd opened this issue Jan 5, 2022 · 8 comments
Open

how to prepear my dataset? #15

ccsvd opened this issue Jan 5, 2022 · 8 comments

Comments

@ccsvd
Copy link

ccsvd commented Jan 5, 2022

hi,
thanks for your good job!
but if i want to train my data? how to process?
i see the llff data have cameras.bin images.bin,points3D.bin。。。how to get these?
could you share the code for that?
Thanks.

@vincentfung13
Copy link
Owner

You will need multi-view images in order to train MINE, this includes camera parameters for each image in the scene, as well as a sparse point cloud of the scene for scale calibration in the case that the camera parameters are estimated with structure-from-motion.

Sometimes some of the parameters are proveded, for example RealEstate10K provideds the camera intrinsics and extrinsics, but since they are estimated with SfM, you will need to run a triangulation to generate the point clouds, you can take a look at the point_triangulator interface of colmap: https://colmap.github.io/cli.html

If you are starting from scratch, you can automatically estimate all of the parameters with automatic_reconstructor, which will generate all the bin files in your question.

Hope this helps.

Zijian

@ccsvd
Copy link
Author

ccsvd commented Jan 5, 2022 via email

@ccsvd
Copy link
Author

ccsvd commented Jan 6, 2022

ok,i will check and try!
well,i have some new questions:
1、could you share the training.log file for llff model train?so i can check my train is ok.
2、what is mean the param img_pre_downsample_ratio?it is fixed for any dataset?
3、the llff image size is 512x284, it has to be the same as mode input size? or it can be any size in my new dataset?
thanks for reply!

@anuraguppuluri
Copy link

You will need multi-view images in order to train MINE, this includes camera parameters for each image in the scene, as well as a sparse point cloud of the scene for scale calibration in the case that the camera parameters are estimated with structure-from-motion.

Sometimes some of the parameters are proveded, for example RealEstate10K provideds the camera intrinsics and extrinsics, but since they are estimated with SfM, you will need to run a triangulation to generate the point clouds, you can take a look at the point_triangulator interface of colmap: https://colmap.github.io/cli.html

If you are starting from scratch, you can automatically estimate all of the parameters with automatic_reconstructor, which will generate all the bin files in your question.

Hope this helps.

Zijian

Are two views per scene sufficient for training MINE on thousands of scenes at a time?

@ccsvd
Copy link
Author

ccsvd commented Jan 28, 2022 via email

1 similar comment
@ccsvd
Copy link
Author

ccsvd commented Jan 17, 2023 via email

@tedyhabtegebrial
Copy link

@vincentfung13 Do you have any suggestions on training without SFM, incase if we have a good estimate of the scene's near and far extent, already ?

@ccsvd
Copy link
Author

ccsvd commented Feb 21, 2023 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants