Git Product home page Git Product logo

grouprcnn's People

Contributors

jshilong avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

grouprcnn's Issues

The code

Nice work! However, the code seems not complete.​ For example, in the paper, you use retinaHead to replace RPNHead, but, you do not offer the fixed retinaHead. Hope you can update those incomplete code. Thanks!​

Why the performance of teacher model is so high?

The GroupRCNN is based on the Cascade RCNN.

The performance of Cascade RCNN with 100% label is about 41AP.
image

But the performance of GroupRCNN with 10% label is about 39.2AP, which seems too high.
image

For the PointDETR, the performance of teacher model with 10% label is 23.7AP, which is much lower than yours. But the gap of student performance is not as significant (teacher:39.2 vs. 23.7, student: 32.6 vs. 30.3). It seems very strange.
image

sampling rule for the fully-labeled dataset

Thank you for sharing your great work, GroupRCNN for WSSOD.

As in the below figure, GroupRCNN is trained with a subset (10% or 20% or 50%) of COCO fully-labeled datasets.
image

Are there any rules or algorithms to sample a subset of COCO fully-labeled datasets?
For example, when randomly sampling 10% of fully-labeled datasets, the minor (long-tailed) classes are rarely included in the 10% sampled dataset.

Did you consider the distribution of categories when sampling the COCO dataset or do you have any special rules for sampling the dataset?

Use 2 GPUs

Sorry to bother you, but how to set the file parameters when using two Gpus or one GPU? Is it set directly through the command? When I tried this method, I got an error as shown in the figure below:
image
Hope to receive your reply, thank you!

the result have a wave?

hello, I use the dataset without segmatation and olny bbox information,but the result have a wave about max-min =4MAP.Although I change the point to in the object,the result also wave.Do you have the same? PS:the result only the point to bbox generator.

Pseudo-box prediction

Hi,
Thanks for sharing the code.
Can you clarify if the GroupRCNN approach ensures that a pseudo box is predicted for every point annotation? Can there be points for which no pseudo-box is predicted by the teacher model?

About using myself dataset

I'm sorry to bother you. I encountered a problem: Because my own data set did not have segmentation data set information, I obtained performance A after generating point annotations with box information, and then obtained performance B after manual labeling training, but performance B was less than performance A. My understanding is that manually annotated point annotations are higher than randomly generated point annotations, but the model performance has not been improved, may I ask what the possible problem is?
Looking forward to hearing from you!

about prior

Sorry to interrupt, regarding the program of the prior part, I can understand that it is to get the index of the first three points through the feature map, and get the anchor points through the index to change the template of the generated anchor box to achieve the purpose (the target of the box is consistent with the target of the point annotation). My understanding: the code group_rcnn.py> _get_dummy_bboxes_single: bboxes=self.rpn_head.bbox_coder.decode(priors, bbox_pred, max_shape=img_shape), where priors is [x, y, x,y]+base_anchors, The first part is the anchor point obtained according to the index, the second part is the base anchor box obtained according to the anchor box generator with the anchor point as (0, 0), so that a new anchor box template is established with the information of the point annotation. The proposal (without NMS and TOPK versions) is obtained by calculating the delta2bbox algorithm with bbox_pred (regression parameter obtained using rpn_head) and prior in decode.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.