Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to improve image quality by monosdf rendering #77

Open
small-zeng opened this issue Aug 16, 2023 · 7 comments
Open

How to improve image quality by monosdf rendering #77

small-zeng opened this issue Aug 16, 2023 · 7 comments

Comments

@small-zeng
Copy link

have a multi-room scenario, using 400 images for reconstruction with MonoSDF. The rendered new viewpoints only achieve a PSNR of 21. How can I improve this?

image
image

@small-zeng
Copy link
Author

Do I need to adjust the number of layers in the rendering network or the number of sampling points in the configuration file?

@niujinshuchong
Copy link
Member

Hi, the mesh looks reasonable. Did you use per_image_code in your training?

@small-zeng
Copy link
Author

Hi, the mesh looks reasonable. Did you use per_image_code in your training?
Thank you for your response. I disabled the per_image_code during my training because it seemed ineffective; it was merely an input of image indices. However, my test set undergoes a separate process of random rendering. Would this have an impact on the results? What is the actual role of per_image_code and how does it function?

@small-zeng
Copy link
Author

    if self.per_image_code:
        image_code = self.embeddings[indices].expand(rendering_input.shape[0], -1)
        rendering_input = torch.cat([rendering_input, image_code], dim=-1)
        
    x = rendering_input      

The "image_code" here seems to be just an index input, which would only improve the training views and not enhance the test views.

@niujinshuchong
Copy link
Member

Hi, the per-image-code is proposed in nerf-in-the-wild paper can could model large appearance variance. It's true that it can't improve over test view since we don't have the per-image-code for the test views.

@small-zeng
Copy link
Author

Hi, the per-image-code is proposed in nerf-in-the-wild paper can could model large appearance variance. It's true that it can't improve over test view since we don't have the per-image-code for the test views.

Thank you, are there limitations when using larger multi-room scenes, such as network forgetting issues? How should we go about solving this issue?

@niujinshuchong
Copy link
Member

Hi, in this repo, we sample rays from a single image at each iteration since we use monocular depth loss and the rays should come from the same image. If the scene is big, the model might have forgetting issues. Might be better to adapt it to using rays from multiple images e.g. 16.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants