Git Product home page Git Product logo

depthcov's Introduction

DepthCov

Code for the paper "Learning a Depth Covariance Function" including the training pipeline and a real-time dense monocular visual odometry pipeline DepthCov-VO.

Setup

We have included the environment.yml file for the exact environment used, but the full instructions for required packages are below.

Create a new conda environment

conda create -n depth_cov python=3.8.15

Activate the environment via

conda activate depth_cov

Install the following packages:

conda install pytorch torchvision pytorch-cuda=11.7 -c pytorch -c nvidia
conda install matplotlib h5py
pip install pytorch-lightning==1.8.2 opencv-python-headless open3d pyrealsense2 

For the odometry backend, we require the batched SE3 exponential map from Lietorch https://github.com/princeton-vl/lietorch. Clone the repo and run the setup command while in the new conda environment:

git clone --recursive https://github.com/princeton-vl/lietorch.git
cd lietorch
python setup.py install

Lastly, in the virtual environment, go back to the base directory and install

cd ..
pip install -e . 

Executables in "depth_cov/" including the odometry can then be run.

Please note that the "extra_compile_args" in "setup.py" for both LieTorch and DepthCov may need to be modified to match the desired GPU/CUDA setup.

Network Training

For ScanNet, we used the repo https://github.com/jczarnowski/scannet_toolkit to setup training and validation splits. We used the raw depth maps for training.

For the ScanNet dataloader, we require a .txt file where each row is a comma separated list of "RGB_path,depth_path".

For the NYUv2 dataloader, we require a .txt file where each row is the .h5 file containing both the RGB and depth.

Use depth_cov/train.py to jointly train the GP hyperparameters and the UNet parameters.

Depth Completion Examples

depth_comp.mp4

To see how to perform depth completion, see the depth_cov/depth_completion_viz.py example, which shows the output conditional mean and variance given randomply sampled observations.

We also include an interactive demo to visualize the effect of adding new observations and how a covariance matrix looks in depth_cov/active_depth_viz.py. The user may click on the RGB image or the standard deviation visualization to improve the depth estimate.

Bundle Adjustment with Depth Priors

title_bundle_adj.mov

To run the bundle adjustment example depth_cov/bundle_adjustment.py, follow the additional installation instructions at https://github.com/edexheim/DepthCovFactors.

Real-Time Odometry (DepthCov-VO)

dvo3.mp4

The file depth_cov/odom_dataset.py can be used to run the odometry on a dataset while depth_cov/odom_demo.py is set up to run the real-time odometry with an Intel RealSense.

We have tested the odometry on CUDA 11 with an RTX 3080.

Since we have now focused on the real-time aspect, the odometry code has been overhauled since the original paper version. There are now multiple threads and some added features. We include a config file that was originally used for the single threaded odometry with one tracking and one mapping iteration per frame in the paper, but this has not been tuned for the new system and lacks the improved functionality, as we focus on achieving ~30 FPS.

TUM

Download sequences from https://cvg.cit.tum.de/data/datasets/rgbd-dataset/download.

ScanNet

Please refer to the ScanNet Training section on how to download sequences.

Real-time with Intel RealSense

We use the RGB stream for a RealSense camera via the pyrealsense2 API. We assume the factory calibration is sufficient, as we load this directly in the dataloader. The default stream settings we use are 640x480 at 60 FPS, but for USB2 this may not be possible, so adjust realsense.yml as needed.

Citation

If you found this code/work to be useful in your own research, please consider citing the following:

@inproceedings{dexheimer2023depthcov,
  title={Learning a Depth Covariance Function},
  author={Eric Dexheimer and Andrew J. Davison},
  booktitle={IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).},
  year={2023}
}

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.