Git Product home page Git Product logo

driveandsegment's People

Contributors

vobecant avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

driveandsegment's Issues

training code

hi, authors. Thanks for great work!

Will you release the training code for this work

How to project the 'ignored' labels from lidar range image to the rgb image

Hi,

Many thanks for reading my question.

Could you please kindly explain more detailly about how you project the 'ignored' labels from lidar range image to the rgb image. In my opinion, building accurate correspondence between range image and rgb image requires accurate distance measurements, while the distance of the points with 'ignored' labels are invalid.

Thanks again.

Regarding Segment-wise unsupervised labeling

Hi,

Thanks for your novel contribution.

Could you clarify the Segment-wise unsupervised labeling part, i.e., section 3.2?

I want to know how you mask out the pixels belonging to the background or another object within the tight rectangle before feeding it to the Vit model. I have tried setting the pixel value to 255 or 0 but feeding such images to the Vit model does not give an informative CLS token for K-means clustering.

Do you perform preprocessing (like PCA) on the CLS token set before performing mini-batch k-means? Also, please comment on the cluster quality obtained from mini-batch k-means. In my case, the cluster quality is not suitable for further processing.

I really appreciate any help you can provide.
Thanks in advance.

Questions about the datasets

Hi,

Many thanks for sharing your excellent work.

The following are some questions about the configurations of the datasets.

  1. As stated in Sec. 4.1: 'We train our models on about 7k images from the Waymo'. However, as far as I know, the Waymo training set has around 800 sequences with each sequence containing roughly 200 images. Could you please kindly provide more detailed description about how the images are sampled.
  2. Since Waymo dataset contains rgb and range images captured with multiple cameras and lidars(FRONT, TOP, SIDE etc. ). Could you please kindly describe which one you used in the paper.

Many Thanks.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.