Git Product home page Git Product logo

Comments (5)

Wuziyi616 avatar Wuziyi616 commented on July 17, 2024

Btw I have a few questions:

  • Is the teaser gif rendered by the NeRF models you trained? The images look very sharp. The images on the webpage also look good. However, when I go to the hugging face data repo, the previewed images seem to have many floaters and blurry artifacts (e.g. this laptop)
  • Is there a way to select scenes where your pre-trained Plenoxels can generate sharp results? E.g. do you record the PSNR of all scenes you trained? Or if the blurry images are because their rendering views are too different from the camera poses in the training sequence?
  • I want to build a video dataset with high FPS. The original CO3D only provides sub-sampled images (and their original videos are not high FPS either). This is why I find PeRFception useful -- I can densely sample a trajectory of camera poses, and then rendering along this trajectory will give me high FPS video. But to get high-quality videos, I want each rendered frame to be sharp and look natural. Do you have any tips on how to better leverage your pre-trained Plenoxel models? E.g. sampling the camera trajectory along the training sequence? (Also I feel that Plenoxels trained on CO3D-V2 should have fewer artifacts than V1. Is that correct?)

Sorry for texting so many questions. I really appreciate your work, and I believe it will be very helpful to the entire community. Thanks in advance!

from perfception.

Wuziyi616 avatar Wuziyi616 commented on July 17, 2024

@jeongyw12382 any updates here?

from perfception.

jeongyw12382 avatar jeongyw12382 commented on July 17, 2024

Hi. Sorry for being late. All our team members were busy for the CVPR submission. Here are the responses for you questions.

  1. Thanks for the suggestion. We will fix this issue when updating the second version. We are currently generating the dataset.
  2. Thanks for point the typo. We have just adjusted the lambda_tv_color.
  3. It really depends on the scenes' condition. All the teaser images are picked from our generated dataset.
  4. One of the thing we've observed while generating PeRFception is that PSNR is not an almighty metric. We first picked the top 200 rendered images with PSNR, SSIM, and LPIPS. Then. we picked the overlapping scenes.
  5. This should be related to our future extension. First, as you've mentioned, Plenoxel is a great tool for generating high quality videos if two condition holds: the quality of image should be sufficiently great, and estimated camera poses should be accurate. One tip for the former condition is to extend the training step. Because we renderd more than 10K scenes, we could not train sufficiently many iterations. As we've observed in the many scenes, the training was not perfectly done, i.e., validation curve keeps increasing. For the latter condition, you could try for better SfM tools with a stronger recent SOTA matchers. This was done for given camera poses from the official CO3D. But, we are sure that if we utilize stronger camera
    calibration tools, such as SuperGlue, to acquire camera poses, the result should be much better.

Thanks :)

from perfception.

jeongyw12382 avatar jeongyw12382 commented on July 17, 2024

Feel free to reopen issue if you have any questions of helps for this issue. We'll immediately reflect your comment on the upcoming update.

from perfception.

Wuziyi616 avatar Wuziyi616 commented on July 17, 2024

@jeongyw12382 Thanks a lot for your reply! That answers most of my questions. Just a minor one: you are using depth maps to initialize ScanNet NeRF training. Have you tried similar things to CO3D, as it also provides sparse point clouds (though it's reconstructed by COLMAP)? Also, have you tried e.g. depth-supervision loss to improve the NeRF performance on CO3D?

Also, I tried training on CO3D-V2, though the numerical results (e.g. PSNR) clearly improve, the floaters don't seem to improve at all. That's very weird... I'm attaching the 2 rendered videos on the same data from V1 and V2, I really cannot tell the difference

fgbg_nerf-co3d_v1.mp4
fgbg_nerf-co3d_v2.mp4

from perfception.

Related Issues (16)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.