Git Product home page Git Product logo

dise-domain-invariant-structure-extraction's People

Contributors

a514514772 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar

dise-domain-invariant-structure-extraction's Issues

FCN8 code

Hi

I want to try this out with FCN 8 with VGG16 backend. Do you have some code to show how the decoder architecture changes and how you use the skip connections? If not code , do you have some documentation on the changes to the architecture.

Thanks in advance.

Translation Texture Loss is missing

Hi,

Thanks for sharing your codes.

As per the paper a "Translation Texture Loss" was employed to maintain the texture similarity between the translated images and target images. I couldn't find the use of "Translation Texture Loss" in your code. Can you please comment on it.

Best Regards,
Subeesh.

Unable to reproduce the values in the paper

Hello,

I run the code "train_dise_gta2city.py" following the procedure explained in this project page. The only change I have done was to keep the batch_size as 1, to reduce the memory requirement. I got 38.4% mIoU on val set. This is a big difference as compared to the value of 45.4% reported in the paper. Can you please help me to understand the potential reasons behind this performance drop.

Some of the possible reasons which I could guess are the following.

  1. The mIoU scores are computed at a resolution of 512 x 1024, while the original images are of size 1024 x 2048. In the paper, the resolution used to report the values are not mentioned. May be authors have reported the values at a resolution of 1024 x 2048? Just to check if this is the reason, I used the pretrained weights provided by the authors and got a score of 44.2% for images at resolution 512 x 1024. Therefore, I am assuming that the resolution of test images is not the reason behind performance drop

  2. As per the paper, authors have used pretrained weights from PASCAL VOC dataset to initialise the encoder. This can also be the reason behind performance drop. However, even when I start the training scheme with the pretrained weights provided by the authors, the performance goes down eventually and will start to fluctuate around 38-39%.

Has anyone succeeded to get values around 44 % up on experimentation with this code?

Regards,
Subeesh

The source only results of Cityscapes

Hi. The work is excellent great.

I have downloaded it and trained it on my server (4 GPUs, each GPU with 1 batchsize, 100k iters). I only used the GTA5 data to train the seg network (resnet101) and test on the cityscapes dataset (source only). I found the evaluation results is ~24% mIoU, which have a large gap compared to paper (39%). However, if I change the model from eval() to train() state in the evaluation, I got 34% mIoU. May I ask which state you have used in the paper? If it is eval() state, could you help us to find the potential issues?

Thanks.

Domain classifiers

Hi,

First of all, thanks for sharing your codes.
Secondly, I am wondering why there are two domain classifiers in the codes each optimized separately (dclf1 and dclf2). The domain classifier is T in Figure 2. Right? As far as I understood the pixel level classification (T) is only applied for comparison to the label ground truth of source domain and there is only one classifier. Am I right? I appreciate it if you guide me about it.

Best Regards,
Samane

11GB GPU for training

Hello
I was wondering if there is a way that I can fit the model for training on a 11GB GPU without losing a lot in the segmentation accuracy. I initially tried to reduce the cropped size by half but that reduced the resulting segmentation accuracy by 17%.

Thanks for your help

deeplab optim_parameters not called. same lr rate for all layers

Hi

You have two different learning rates for the shared encoder( deeplab model) in the optim params function in model.py

def optim_parameters(self, learning_rate): return [{'params': self.get_1x_lr_params_NOscale(), 'lr': 1 * learning_rate}, {'params': self.get_10x_lr_params(), 'lr': 10 * learning_rate}]`

But this function is not called during the optimizer initilzation and you load all parameters with one learning rate.

This is also different to the AdaptSeg code.

Is this on purpose ? Is this giving better results than using the seperate learning rates for layer1 to layer4 and a different one for layer5 and layer6

About Source-Only Performance

Hi, congratulation on the work! I noticed that you used PASCAL VOC pretrained model and got source-only performance of 39.8 which is higher than that in AdaptSegNet of 36.6. But in the code of seg_model.py, I see that you still use the pretrained model in AdaptSegNet (the same restore_from). I'm a little confused about this. And is the weights in the google drive you provided pretrained on PASCAL VOC as the paper said? And I wonder if the PASCAL VOC pretrained model is available, thanks.

pytorch version and results reproduction

Hi
Thanks for your code!
I trying to reproduce the result in the paper but i get at best 42.4 miou with a batch size of 2 on a P6000 nvidia card with your default setting.

By the way one's needs a card with at least 20Gb of memory to run your code, i try on a card with 16 Gb and i got a memory issue...

I use pytorch 1.2 and i see that you used pytorch 0.3.1 so maybe it's the reason for the performances degradation as i know there is a difference in the default behaviour in upsample between those 2 versions.
Have you try to re run the code with a newer version of pytorch than the 0.3.1?
Thanks

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.