Comments (3)
Top-1 accuracy on imagenet2012
should be 85.082% for that model.
We use im = tf.image.resize(im, [384, 384])
and then normalize the image m = (im - 127.5) / 127.5
- see https://github.com/google-research/vision_transformer/blob/master/vit_jax/input_pipeline.py
from vision_transformer.
My current B_16 top-1 accuracy is also a bit lower at 83.6%. One difference I found looking at this more closely is that the image normalization I was using was based on the vit_jax.ipynb
sample notebook, which is subtly different:
I've replaced this in my own code now with the m = (im - 127.5) / 127.5
version and am seeing slight differences and will report back if this closes the gap on the reported score.
[Edit: ok - the normalization difference in score is very small (0.01%) - so also checking resize routing on TiT-B_16 top1 score and will report back]
from vision_transformer.
So that normalization difference cited above changes very little (0.01%) and so I looked closer at the resizing code, but matching that exactly also only moved my score about 0.4% - the best top-1 accuracy I could get matching both for the ViT-B_16 model was 84.01%. So I'd been keen to do a deeper dive to diagnose this if anyone else has gotten to 85% and would be willing to compare exact results. I'll suggest a good starting point would be to look at the accuracy of a particular class (50 images each) and then potentially dive down into any differences for particular files, etc.
from vision_transformer.
Related Issues (20)
- Fine-Tuning HOT 3
- ERROR: Could not find a version that satisfies the requirement tensorflow_text (from vit-jax) (from versions: none) ERROR: No matching distribution found for tensorflow_text HOT 1
- If the weights of vit-base trained with dropout available? HOT 1
- How do I download the vit_base_patch8_384.pth
- Package versions' confliction [Windows] HOT 2
- Question about commercial usage of LiT model checkpoints
- Shouldn't accumulate_gradient pass rng_key?
- GPU Requirement to use vision transformer HOT 1
- flax.errors.CallCompactUnboundModuleError
- ViT
- can export the pretrained model to onnx or pytorch? HOT 2
- Vision transformer
- Import error in Jax (colab) HOT 3
- fine-tune imagenet21k_ViT-B_16.npz with pre_logits? HOT 2
- KeyError: 'embedding/kernel is not a file in the archive' HOT 1
- Hyperparameter issues HOT 1
- All attempts to get a Google authentication bearer token failed, returning an empty token.
- Where is ViT-22B?
- Is it possible to provide the checkpoint?
- How can I use imagenet21k+imagenet2012/ViT-B_16.npz to inference
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from vision_transformer.