Comments (12)
Yes, we have compared it in our method and at least in our setting, it gets better performance than OFT. Please take a look at Table 3 in the paper.
from dsgn.
Yes, we have compared it in our method and at least in our setting, it gets better performance than OFT. Please take a look at Table 3 in the paper.
OK, I see. AP_BEV 19.92 ranks 3rd in KITTI BEV leaderboards compared with the other monocular methods. What a good performance. Could you also make this experiment code open sourced? Thanks!
from dsgn.
As mentioned in evluation metric section in paper, the reported results in validation set are evaluated under the old evaluation metric, which seems to (can) be higher than the new metric. This is a old project here. As I remembered, the monocular experiment is ran by simply dropping the right view and projecting the features to BEV directly. The projection code is available in the code. Currently I have other projects at hand. I suggest you try it yourself.
from dsgn.
As mentioned in evluation metric section in paper, the reported results in validation set are evaluated under the old evaluation metric, which seems to (can) be higher than the new metric. This is a old project here. As I remembered, the monocular experiment is ran by simply dropping the right view and projecting the features to BEV directly. The projection code is available in the code. Currently I have other projects at hand. I suggest you try it yourself.
I didn't notice the old metric used in the ablation study. It may decrease about 4% ~ 5% with the new metric. Thanks!
from dsgn.
As mentioned in evluation metric section in paper, the reported results in validation set are evaluated under the old evaluation metric, which seems to (can) be higher than the new metric. This is a old project here. As I remembered, the monocular experiment is ran by simply dropping the right view and projecting the features to BEV directly. The projection code is available in the code. Currently I have other projects at hand. I suggest you try it yourself.
One more questions. How do you apply depth supervision to 3D Volume when there is no Plane Sweep Volume?
from dsgn.
Its implementation is a bit tricky since only the sparse ray is visiable when supervising the voxel grid. You need to project back the depth map to point cloud and ignore the voxels that are not within the projection ray.
from dsgn.
Its implementation is a bit tricky since only the sparse ray is visiable when supervising the voxel grid. You need to project back the depth map to point cloud and ignore the voxels that are not within the projection ray.
I mean we can not apply Softmax to 3D Volume along the z axis like Plane Sweep Volume to get the depth because of the camera frustum.
You apply binary classification to the 3D Volume in occupancy manner. Firstly, projecting the image to point cloud according to sparse depth map like pseudo-lidar does. Secondly, if there is a 3D point, setting the voxel to positive. If there is no 3D point, setting the voxel to negetive. Is it?
from dsgn.
Right. But you need to ignore some voxels as they have no depth.
from dsgn.
Right. But you need to ignore some voxels as they have no depth.
Ok. Only set the label along the sparse lidar rays. Most voxels are ignored. I think it is a little complicate to calculate the 3D Volume's occupancy depth label.
Can I apply some 3D Covs after the 3D Volume and add a depth head? So I can reduce the z axis to get a (H', W') feature map and do depth supervision in this head. How do you think about this implement? I hope the depth supervision can improve the 3D detection performance.
from dsgn.
I think it is OK. Good to know your progress.
from dsgn.
I think it is OK. Good to know your progress.
Thanks for your detail reply.
from dsgn.
I found an interesting paper which well answered my question. https://arxiv.org/pdf/2103.01100.pdf
from dsgn.
Related Issues (20)
- Train, Val, Test split HOT 4
- Inference Speed HOT 2
- 同一个目标多个预测框 HOT 9
- the error of the depth estimation HOT 2
- What is the supervision of point cloud? HOT 6
- may i train this model without the depth estimation? HOT 1
- CUDA Of Memory HOT 11
- Model that has been trained
- tensorflow / tensorflow lite HOT 2
- About the construction of Plane Sweep Volume HOT 4
- where's the split files?
- 作者你好,你的build_cost_volume_forward部分是c++实现的吗? HOT 1
- TypeError: expected dtype object, got 'numpy.dtype[float64]' in eval.py HOT 1
- ./configs/default/config_disp.py这个文件是用来做什么的,是只包含深度估计网络的config吗 HOT 1
- Question about setup(performance issue) HOT 2
- Using 3D Geometry Volume to predict depth. HOT 4
- Results don't match with paper on KITTI Leaderboard HOT 7
- CUDA out of memory HOT 7
- The wrong outputs of the pre-trained model HOT 7
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from dsgn.