Git Product home page Git Product logo

fastpointtransformer's People

Contributors

chrockey avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

fastpointtransformer's Issues

Questions about model runs memory and FLOPs

Hello, I would like to ask you how much memory and FLOPs your model runs on S3DIS, I would like to quote your paper, but there is something wrong with my computer, so I thought I would ask you directly!

[Question] hardware used and training time

Hi

The paper mentions an inference time extremely reduced compared to the original PointTransformer, but I was also curious about the time it took to train the model. What GPUs did you use and how much time did it take to train compared to PointTransformer ?

Thx a lot

Questions about fast point transformer

Thanks for your amazing work, and I have few questions about the implemention of this architecture. Thank you for any answers.

  1. In the code of LightweightSelfAttentionLayer, why the inter postion embeding is initialized as a learnable random variable ?

self.inter_pos_enc = nn.Parameter(torch.FloatTensor(self.kernel_volume, self.num_heads, self.attn_channels))
nn.init.normal_(self.inter_pos_enc, 0, 1)

According to Fig 3 in the paper, shouldn't it be obtained from the coordinate difference between the current voxel and neighboring voxels ?

  1. How many specific neighboring voxels are indexed in LightweightSelfAttentionLayer ?Is the number of neighboring voxels determined by kernel_size in the input parameter? Is the neighboring voxels the valid voxels contained in the kernel ?

Could it utilize the inter-frame information?

Hi,Thanks for your great work first,but I have some questions about the attention block.
For example, I set batch size = 2, how can i find the query voxels around both two frames.

About the environment

Could you share some environmental information? I can successfully install pytorch according to setup.py, but there is a conflict when I install openblas-devel.

import cuda_sparse_ops

Thanks to the author for such quality code. I have some questions for the author,
in the sparse_ops.py file.
import cuda_sparse_ops
keeps reporting errors.
How to solve it? ? ?
And how to run setup.py in src.cuda_ops??????

How to get the access to wandb?

An error is reported when training the dataset, saying "Error while calling W&B API: permission denied (<Response [403]>)". I checked the documentation of wandb and it says I don't have the project permissions. I want to know how to get the permission, thanks for your help!

how to download the dataset?

Before calling the script preprocess_s3dis.py,I need to download the data manually, right? Or automatic download in the script preprocess_s3dis.py? If I need to download it automatically,How to download the dataset?

The stride parameter in LSA layer

Hello, I 'd like to know about whether in the implementation, the stride shoud be kept as 1. What other things I may need to do if I want
to expand it to an arbitrary number.

questions

Thanks for the great work! I have some questions.

  1. I wonder where the GPU memory budget cost in FastTrans. Because I only have GTX1080, 12G. Do you test the model for small-scale task like ShapeNet DataSet? And can I apply the lightweight self-attention to get a better feature embedding for ShapeNet classification in my 1080 machine? Maybe I can use a smaller FastTrans?
  2. I notice that the voxel size is used for data augmention. If i use the model for ShapeNet classification which normalized coordinates lie between -1 and 1. Can I remove the data augmention and just use original coordinates? Or how to get a feasible voxel size ?

How to get access to wandb?

It occurs when I start to train on S3DIS datset via the command "python train.py config/s3dis/train_fpt.gin"
78186242395619
I tried installing the module using pip but it said "No matching distribution found for cuda_sparse_ops", and I didn't find any solution on the Internet. Is it because there was something wrong with my installation?

RuntimeError

result, COO, vals = MEB.coo_spmm_average_int32(
RuntimeError: at /FastPointTransformer/thirdparty/MinkowskiEngine/src/gpu.cu:100
I encountered this issue during runtime.I don't know how to solve it. Can you help me take a look

DDP/DP training - multigpu

Hi @chrockey, great work!

Can you guide me on how to set up multigpu training? I have only 20GB gpus available, and when using batch size of 2 I obtain poor performance (~6% lower mIoU and mAcc; probably due to the batch norm and batch size).

If I add multigpu support (DDP) according to the example from the ME repository the learning is blocked, i.e. it never starts.

Any help will be appreciated. You commented "multi-GPU training is currently not supported" in the code. Have you had similar issues as I mentioned?

Thanks!

Why δ(vi-vj) is O(KD)?

Hi!
When I read your paper at reducing space complexity section, it says that the space complexity of δrel(vi-vj) is O(KD),I can't understand it.
I think there are K neighbors for each voxel,why not O(IKD)? I hope you can help me ,Thanks!

Training Log

Hellow! Thank you for your awesome work!
I find that it takes 20 hours on A100 to train your model. Could I have a look at your training log on S3DIS?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.