nianticlabs / diffusionerf Goto Github PK
View Code? Open in Web Editor NEW[CVPR 2023] DiffusioNeRF: Regularizing Neural Radiance Fields with Denoising Diffusion Models
License: MIT License
[CVPR 2023] DiffusioNeRF: Regularizing Neural Radiance Fields with Denoising Diffusion Models
License: MIT License
This is a cool paper, but I would like to ask what GPU is needed if I want to test.
Dear authors,
I read your amazing paper.
When will you release a supplementary?
Thank you!
Thanks for the paper. I'm trying to run the code on LLFF/room data as the author's. But I got an Error described as follows.
Skipping scene - not included in specified scenes ['room']
Skipping scene - not included in specified scenes ['room']
Skipping scene - not included in specified scenes ['room']
Skipping scene - not included in specified scenes ['room']
Skipping scene - not included in specified scenes ['room']
Skipping scene - not included in specified scenes ['room']
Scene data/nerf_llff_data/room
Using bound 7.5 and scale 1.5
Final bound and scale: 7.5 1.5
Fitting to scene data/nerf_llff_data/room
main.nerf running with options Namespace(path=PosixPath('data/nerf_llff_data/room'), test=False, workspace=PosixPath('runs/example/3_poses/room'), seed=0, iters=10000, lr=0.01, ckpt='latest', num_rays=1024, num_steps=512, upsample_steps=128, max_ray_batch=1024, downsample_val=8, downsample_train=8, test_mode=['test_eval'], eval_interval=5, max_val_imgs=None, num_train_poses=3, patch_regulariser_path='models/rgbd-patch-diffusion.pt', patch_sample_downscale_factor=4, patch_weight_start=0.2, patch_weight_finish=0.2, patch_reg_start_step=0, patch_reg_finish_step=2500, initial_diffusion_time=0.1, normalise_diffusion_losses=True, apply_geom_reg_to_patches=True, spread_loss_strength=1.5e-05, seg_loss_strength=1e-06, weights_sum_loss_strength=0.001, net_l2_loss_strength=1e-07, reg_ramp_start_step=3000, reg_ramp_finish_step=8000, use_frustum_regulariser=True, frustum_check_patches=False, frustum_regularise_patches=True, frustum_reg_initial_weight=1.0, frustum_reg_final_weight=0.01, mode='llff', color_space='srgb', preload=False, bound=7.5, scale=1.5, min_near=0.2, density_thresh=10, error_map=False, clip_text='', rand_pose=-1, only_run_on=['room'], num_train=[3], force=False, normalise_length_scales_to=None, normalise_length_scales_to_at_least=7.5)
Building foreground model
Made LLFF-style train set of 35 images and test set of 6 images
Frames now has len 3
Images: ['m/images/DJI_20200226_143851_396.JPG', 'm/images/DJI_20200226_143918_576.JPG', 'm/images/DJI_20200226_143946_704.JPG']
Loading trainval data:: 0%| | 0/3 [00:00<?, ?it/s][ WARN:[email protected]] global /home/conda/feedstock_root/build_artifacts/libopencv_1671461531762/work/modules/imgcodecs/src/loadsave.cpp (239) findDecoder imread_('data/nerf_llff_data/room/m/images/DJI_20200226_143851_396.JPG'): can't open/read file: check file path/integrity
Loading trainval data:: 0%| | 0/3 [00:00<?, ?it/s]
Traceback (most recent call last):
File "/root/miniconda3/envs/diffusionerf/lib/python3.9/runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/root/miniconda3/envs/diffusionerf/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/root/autodl-tmp/diffusionerf-main/nerf/evaluate.py", line 114, in
main(opts)
File "/root/autodl-tmp/diffusionerf-main/nerf/evaluate.py", line 100, in main
run(opts_tmp)
File "/root/autodl-tmp/diffusionerf-main/main_nerf.py", line 44, in run
train_set = NeRFDataset(opt, device=device, type='trainval', downscale=opt.downsample_train)
File "/root/autodl-tmp/diffusionerf-main/nerf/provider.py", line 232, in init
assert image is not None, f'Unable to find image at {f_path}'
AssertionError: Unable to find image at data/nerf_llff_data/room/m/images/DJI_20200226_143851_396.JPG
The path is always wrong and as you can see it adds the last letter of the dataset class as path , so it triggers the Assertion Error.
How can I solve the problem? Thanks for your patience and resolution.
Thanks for the cool paper. I'm trying to reimplement the local patch-based diffusion model. I have some questions
I assume that you concatenate the input RGBD?
How do you normalize the depth from hypersim? The depth is in meters? and I assume you somehow normalize to the range [-1,1], a similar range as the RGB values? Also, how do you normalize the depth from the rendering nerfs, such that it is in the range [-1, 1]?
What scheduler (DDIM? DDPM?) do you for the reverse process? Figure 3 c?
What architecture do you use for the diffusion model? I'm currently implementing it in huggingface like this (bottleneck resolution 6x6).
model = UNet2DModel(
sample_size=48, # the target image resolution
in_channels=len(config.input), # the number of input channels, 3 for RGB images
out_channels=len(config.input), # the number of output channels
layers_per_block=2, # how many ResNet layers to use per UNet block
block_out_channels=(64, 128, 256, 512),
down_block_types=(
"DownBlock2D",
"DownBlock2D",
"AttnDownBlock2D",
"DownBlock2D",
),
up_block_types=(
"UpBlock2D",
"AttnUpBlock2D",
"UpBlock2D",
"UpBlock2D",
),
)
Hi Daniyar,
Thanks for your excellent work! Here I have some questions about the DDM prior.
As you mentioned in the main paper, the DDM is trained on RGBD patches of Hypersim dataset, to model a prior over the distribution of RGBD patches. During training, the gradients from the DDM is propagated to both of the depth and color of the rendered patches with different weights.
I wonder how do you formulate the DDM? Does the DDM simply take as inputs the concatenation of RGB & D in the channel dimension? or using a depth2img DDM to make the gradient propagate through the conditional branch? Can you share your code or supplementary material?
Thanks in advance!
I found that the mesh extracted by NGP cannot align with the GT of the DTU dataset, it doesn't work even after using Monosdf's culling. I would like to know how you handled it. If possible, could you please upload the code?
Hello,
thank you for sharing the code of your great work.
I noticed that in the paper, the comparison with other sota methods shows inconsistencies between lpips and other metrics where your method is better at lpips but not necessarily on the others, especially on the extreme setting of 3 views in DTU dataset.
I see in your code that you load both lpips_alex and lpips_vgg models for evaluation, but it is lpips_alex that is being used, so I wonder if it was used for the comparison in the paper as well.
It is worth noting that the methods that are being compared to, such as PixelNerf and RegNerf use lpips vgg for their evaluation.
#!/bin/bash
mkdir ./runs/
python -m main_nerf /data3/liuyu/nerf_llff_data/flower/
--mode blender
--upsample_steps 128
--num_rays 1024
--max_ray_batch 1024
--mode llff
--downsample_val 8
--reg_ramp_start_step 3000
--reg_ramp_finish_step 8000
--iters 10000
--min_near 0.2
--spread_loss_strength 1.5e-5
--num_train 3
--use_frustum_reg
--patch_weight_start 0.2
--patch_weight_finish 0.2
--normalise_diffusion_losses
--patch_regulariser_path models/rgbd-patch-diffusion.pt
--frustum_regularise_patches
--apply_geom_reg_to_patches
--downsample_train 8
--patch_reg_start_step 0
--workspace ./runs/train/3_poses/flower/
I used this command and the following error occurred in the flower scene of training llff.:
==> Start Training Epoch 1, lr=0.010000 ...
0% 0/201 [00:00<?, ?it/s]Traceback (most recent call last):
File "/data1/liuyu/miniconda3/envs/diffusionerf/lib/python3.9/runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/data1/liuyu/miniconda3/envs/diffusionerf/lib/python3.9/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/data1/liuyu/MPI/diffusionerf/main_nerf.py", line 145, in
run(opt)
File "/data1/liuyu/MPI/diffusionerf/main_nerf.py", line 122, in run
trainer.train(train_loader, valid_loader, max_epoch)
File "/data1/liuyu/MPI/diffusionerf/nerf/utils.py", line 657, in train
self.train_one_epoch(train_loader)
File "/data1/liuyu/MPI/diffusionerf/nerf/utils.py", line 776, in train_one_epoch
preds, truths, loss, extra_outputs = self.train_step(data)
File "/data1/liuyu/MPI/diffusionerf/nerf/utils.py", line 412, in train_step
outputs = self.model.render(rays_o, rays_d, staged=False, bg_color=bg_color, perturb=True, force_all_rays=False, **vars(self.opt))
File "/data1/liuyu/MPI/diffusionerf/nerf/renderer.py", line 370, in render
results = _run(rays_o, rays_d, **kwargs)
File "/data1/liuyu/MPI/diffusionerf/nerf/renderer.py", line 129, in run
z_vals = perturb_z_vals(z_vals)
File "/data1/liuyu/MPI/diffusionerf/nerf/rendering_helpers.py", line 31, in perturb_z_vals
assert (deltas >= 0).all()
AssertionError
0% 0/201 [00:00<?, ?it/s]
Hi,
thanks for the excellent work, but it is really weird that your gradient of RGBD data over (RGB,d) added into the gradient of photometric loss over(σ,c), can you please help to elaborate on it?
Hi,
Thanks for your amazing paper.
Can you please share details about the training of the RGBD DDM?
How many patches did you use? how long did it train?
Thank you!
thanks for your excellent work, i just use 'bash scripts/run_diffusionerf_example.sh', however, the result is not good , just about psnr 12, i see the predict img is totally blank , are there lack of some weights files? thanks~
I have cloned diffusionerf and tried to execute conda env create -f environment.yml
Commented out the line - git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
from environment.yml file, and tried creating the environment, it worked.
Tried to pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
seperately and got this error:
pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
Collecting git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
Cloning https://github.com/NVlabs/tiny-cuda-nn/ to /tmp/pip-req-build-tfx6b0pr
Running command git clone --filter=blob:none --quiet https://github.com/NVlabs/tiny-cuda-nn/ /tmp/pip-req-build-tfx6b0pr
Resolved https://github.com/NVlabs/tiny-cuda-nn/ to commit 6f018a9cd1b369bcb247e1d539968db8e48b2b3f
Running command git submodule update --init --recursive -q
Preparing metadata (setup.py) ... done
Building wheels for collected packages: tinycudann
Building wheel for tinycudann (setup.py) ... error
error: subprocess-exited-with-error
× python setup.py bdist_wheel did not run successfully.
│ exit code: 1
╰─> [215 lines of output]
/tmp/pip-req-build-tfx6b0pr/bindings/torch/setup.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html
from pkg_resources import parse_version
Building PyTorch extension for tiny-cuda-nn version 1.7
Obtained compute capability 86 from PyTorch
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2021 NVIDIA Corporation
Built on Thu_Nov_18_09:45:30_PST_2021
Cuda compilation tools, release 11.5, V11.5.119
Build cuda_11.5.r11.5/compiler.30672275_0
Detected CUDA version 11.5
Targeting C++ standard 17
running bdist_wheel
running build
running build_py
creating build
creating build/lib.linux-x86_64-cpython-39
creating build/lib.linux-x86_64-cpython-39/tinycudann
copying tinycudann/__init__.py -> build/lib.linux-x86_64-cpython-39/tinycudann
copying tinycudann/modules.py -> build/lib.linux-x86_64-cpython-39/tinycudann
running egg_info
creating tinycudann.egg-info
writing tinycudann.egg-info/PKG-INFO
writing dependency_links to tinycudann.egg-info/dependency_links.txt
writing top-level names to tinycudann.egg-info/top_level.txt
writing manifest file 'tinycudann.egg-info/SOURCES.txt'
reading manifest file 'tinycudann.egg-info/SOURCES.txt'
writing manifest file 'tinycudann.egg-info/SOURCES.txt'
copying tinycudann/bindings.cpp -> build/lib.linux-x86_64-cpython-39/tinycudann
running build_ext
/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py:813: UserWarning: The detected CUDA version (11.5) has a minor version mismatch with the version that was used to compile PyTorch (11.6). Most likely this shouldn't be a problem.
warnings.warn(CUDA_MISMATCH_WARN.format(cuda_str_version, torch.version.cuda))
/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py:820: UserWarning: There are no g++ version bounds defined for CUDA version 11.5
warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}')
building 'tinycudann_bindings._86_C' extension
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt/src
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/src
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/build/temp.linux-x86_64-cpython-39
creating /tmp/pip-req-build-tfx6b0pr/bindings/torch/build/temp.linux-x86_64-cpython-39/tinycudann
Emitting ninja build file /tmp/pip-req-build-tfx6b0pr/bindings/torch/build/temp.linux-x86_64-cpython-39/build.ninja...
Compiling objects...
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)
[1/10] c++ -MMD -MF /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt/src/os.o.d -pthread -B /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/dependencies/fmt/src/os.cc -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt/src/os.o -std=c++17 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
[2/10] c++ -MMD -MF /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt/src/format.o.d -pthread -B /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/dependencies/fmt/src/format.cc -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/dependencies/fmt/src/format.o -std=c++17 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
[3/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/object.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/object.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/object.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/object.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/object.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[4/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/network.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/network.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/network.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/network.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/network.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[5/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/common_host.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/common_host.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/common_host.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/common_host.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/common_host.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[6/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/cpp_api.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cpp_api.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cpp_api.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/cpp_api.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cpp_api.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[7/10] c++ -MMD -MF /tmp/pip-req-build-tfx6b0pr/bindings/torch/build/temp.linux-x86_64-cpython-39/tinycudann/bindings.o.d -pthread -B /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -O2 -isystem /home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include -fPIC -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/bindings/torch/tinycudann/bindings.cpp -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/build/temp.linux-x86_64-cpython-39/tinycudann/bindings.o -std=c++17 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
[8/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/encoding.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/encoding.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/encoding.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/encoding.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/encoding.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[9/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/cutlass_mlp.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cutlass_mlp.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cutlass_mlp.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/cutlass_mlp.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/cutlass_mlp.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
[10/10] /usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/fully_fused_mlp.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/fully_fused_mlp.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
FAILED: /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/fully_fused_mlp.o
/usr/bin/nvcc -I/tmp/pip-req-build-tfx6b0pr/include -I/tmp/pip-req-build-tfx6b0pr/dependencies -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/cutlass/tools/util/include -I/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/torch/csrc/api/include -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/TH -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/include/THC -I/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/include/python3.9 -c -c /tmp/pip-req-build-tfx6b0pr/src/fully_fused_mlp.cu -o /tmp/pip-req-build-tfx6b0pr/bindings/torch/src/fully_fused_mlp.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -std=c++17 --extended-lambda --expt-relaxed-constexpr -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_HALF2_OPERATORS__ -Xcompiler=-Wno-float-conversion -Xcompiler=-fno-strict-aliasing -gencode=arch=compute_86,code=compute_86 -gencode=arch=compute_86,code=sm_86 -DTCNN_PARAMS_UNALIGNED -DTCNN_MIN_GPU_ARCH=86 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=_86_C -D_GLIBCXX_USE_CXX11_ABI=0
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/tmp/pip-req-build-tfx6b0pr/dependencies/fmt/include/fmt/core.h(288): warning #1675-D: unrecognized GCC pragma
/usr/include/c++/11/bits/std_function.h:435:145: error: parameter packs not expanded with ‘...’:
435 | function(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:435:145: note: ‘_ArgTypes’
/usr/include/c++/11/bits/std_function.h:530:146: error: parameter packs not expanded with ‘...’:
530 | operator=(_Functor&& __f)
| ^
/usr/include/c++/11/bits/std_function.h:530:146: note: ‘_ArgTypes’
ninja: build stopped: subcommand failed.
Traceback (most recent call last):
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1808, in _run_ninja_build
subprocess.run(
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/subprocess.py", line 528, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "<string>", line 2, in <module>
File "<pip-setuptools-caller>", line 34, in <module>
File "/tmp/pip-req-build-tfx6b0pr/bindings/torch/setup.py", line 189, in <module>
setup(
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/__init__.py", line 107, in setup
return distutils.core.setup(**attrs)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/dist.py", line 1233, in run_command
super().run_command(command)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/wheel/bdist_wheel.py", line 349, in run
self.run_command("build")
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/dist.py", line 1233, in run_command
super().run_command(command)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/command/build.py", line 131, in run
self.run_command(cmd_name)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/dist.py", line 1233, in run_command
super().run_command(command)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 88, in run
_build_ext.run(self)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 765, in build_extensions
build_ext.build_extensions(self)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 467, in build_extensions
self._build_extensions_serial()
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 493, in _build_extensions_serial
self.build_extension(ext)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 249, in build_extension
_build_ext.build_extension(self, ext)
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 548, in build_extension
objects = self.compiler.compile(
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 586, in unix_wrap_ninja_compile
_write_ninja_file_and_compile_objects(
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1487, in _write_ninja_file_and_compile_objects
_run_ninja_build(
File "/home/harshamupparaju/Softwares/Anaconda/anaconda3/envs/diffusionerf/lib/python3.9/site-packages/torch/utils/cpp_extension.py", line 1824, in _run_ninja_build
raise RuntimeError(message) from e
RuntimeError: Error compiling objects for extension
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for tinycudann
Running setup.py clean for tinycudann
Failed to build tinycudann
ERROR: Could not build wheels for tinycudann, which is required to install pyproject.toml-based projects
System settings:
Ubuntu 22.04, gcc 11.4
Output of nvidia-smi
:
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.86.05 Driver Version: 535.86.05 CUDA Version: 12.2 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 NVIDIA RTX A4000 Off | 00000000:02:00.0 On | Off |
| 86% 92C P2 139W / 140W | 8753MiB / 16376MiB | 94% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
Output of nvcc --version:
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2021 NVIDIA Corporation
Built on Thu_Nov_18_09:45:30_PST_2021
Cuda compilation tools, release 11.5, V11.5.119
Build cuda_11.5.r11.5/compiler.30672275_0
Any help to resolve this issue?
Hello! I'm a bit confused why LLFF_DEFAULT_PSEUDO_INTRINSICS is used instead of ground truth intrinsics.
Hi,
This work is very interesting!
Are you planning to release the code? If yes, what is the approximate timeline for that?
Thanks a lot!
Hi there,
Thanks for your excellent work. Do you mind sharing the DTU results somewhere or the shell script that can reproduce DTU results? I'm trying to reproduce the results on DTU but not sure if I have the correct setting.
Sorry but I am new to this field. I read your paper and was really impressed. I wanted to use it on my own data. How can I create transform.json file and use your network to train my data. Also can I do it with just 2 views of a scene?
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.