Git Product home page Git Product logo

imet's Introduction

iMet Collection 2019 - FGVC6

1st place

  • RandomCropIfNeeded
  • batch accumulation
  • With SIZE = 320 for SEResNext101 / SENet154 and SIZE = 331 for PNasNet-5
  • Filtering predictions by high loss
  • Pseudo labeling add the most confident predictions (highest np.mean(np.abs(probabilities - 0.5)) ) to the training dataset.
  • Culture and tags separately
  • 2nd stage I construct the binary classification dataset: I took 1103 (number of classes) rows per each image and trying to predict that this class relates to this image (0 or 1)
  • feature engineering
  • probabilities of each models, sum / division / multiplication of each pair / triple / .. of models
  • mean / median / std / max / min of each channel
  • brightness / colorness of each image (you can say me that NN can easily detect it — yes, but here i can do it without cropping and resizing — it is less noisy)
  • Max side size and binary flag — height more than width or no (it is a little bit better for tree boosting than just height + width in case of lower side == 300)
  • add all 1000 (number of ImageNet classes) predictions to this dataset
  • lightgbm
  • postprocess
  • (Done)Different threshold for cultures and tags models

2nd place

  • The technique was quite similar to my quickdraw solution. I put top 30 tags and top 20 cultures to the dataset for LGMB

4th place

  • Tag Relevance Prediction

6th place

  • (Done)adjust the threshold for each image according to the max probability of that image
  • Mixup and RandomErasing

9th place

  • knowledge Distillation
  • train only fresh params

10th place

  • Small modifications were made to the network, our logit takes input from the last two layers rather than the last one layer

others

  • two a fully connected layers, one for the culture and one for the tag

Usage

Train

Make folds

python make_folds.py --n-folds 40

Train se_resnext101 from fold 0 to 9:

python main.py train model_se101_{fold} --model se_resnext101_32x4d --fold {fold} --n-epochs 40 --batch-size 32 --workers 8

Train inceptionresnetv2 from fold 5 to 9:

python main.py train model_inres2_{fold} --model inceptionresnetv2 --fold {fold} --n-epochs 40 --batch-size 32 --workers 8

Train pnas models from fold 0 to 4:

python main.py train model_pnas_{fold} --model pnasnet5large --fold {fold} --n-epochs 40 --batch-size 24 --workers 8

Test

The ensemble of these model is used to predict results in imet-predict-final.ipynb.

imet's People

Contributors

yu1ut avatar bamps53 avatar

Watchers

James Cloos avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.