Comments (17)
Interesting. Has this happened more than once? (Depending on random seed, I find that even AlexNet occasionally doesn't learn.)
The next place I'd look is to make sure that the training data is sane,
from squeezenet.
I think the train val lmdb dataset is ok
240G ilsvrc12_train_lmdb/
9.4G ilsvrc12_val_lmdb/
I use create_imagenet.sh
to create these files.
I will train the network again using a different seed.
Thanks.
from squeezenet.
I'm facing the same problem to train SqueezeNet on Darknet... network loss got stuck 5~6 after 80k iterations.
Detection **Avg IOU: 0.354223**, Pos Cat: 0.995564, All Cat: 0.995564, Pos Obj: 0.010827, Any Obj: 0.004997, count: 23
Detection **Avg IOU: 0.352646**, Pos Cat: 0.995592, All Cat: 0.995592, Pos Obj: 0.010121, Any Obj: 0.004997, count: 23
Detection **Avg IOU: 0.368458**, Pos Cat: 0.996199, All Cat: 0.996199, Pos Obj: 0.013484, Any Obj: 0.004997, count: 31
Detection **Avg IOU: 0.384327**, Pos Cat: 0.995394, All Cat: 0.995394, Pos Obj: 0.010156, Any Obj: 0.004997, count: 27
18488: 4.462246, **5.304660 avg**, 0.026577 rate, 4.329206 seconds, 1183232 images
Loaded: 0.000039 seconds
I my case my dataset is converging on vanilla AlexNet, but not converging with SqueezeNet no matters how many times I start training from scratch. Note: Darknet doesn't implement xavier initialization, so I'm using default random initialization.
SqueezeNet v1.1 port for Darknet:
https://gist.github.com/Grabber/65760c4b4e5b4cf9a82f11193a8154dd
@forresti do you have any insight on it? Please answer my messages on LinkedIn or hang up the phone ;)
from squeezenet.
I use caffe, and comment out the random_seed: 42
in solver.prototxt.
Now SqueezeNet v1.1 works very well.
Here is my training log https://gist.github.com/kli-nlpr/5f54a24a1215af9fcd9faaf16be6d54d
from squeezenet.
Don't have more ideas to overcome loss being stuck at 5~6 on Darknet.
from squeezenet.
I met the same problem! When train with learning rate within [0.1, 0.01], loss always stay in 6.9...
then i use learning rate in 0.001, loss decrease well.
i don't know whether it world be overfitting or any other problems when train begin in 0.001 ?
from squeezenet.
Could uou share your solver.txt?
On Wednesday, 27 July 2016, blue [email protected] wrote:
I met the same problem! When train with learning rate within [0.1, 0.01],
loss always stay in 6.9...
then i use learning rate in 0.001, loss decrease well.
i don't know whether it world be overfitting or any other problems when
train begin in 0.001 ?—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
#18 (comment),
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAA9c3LqiULedIipe3YxR222w-jHrk92ks5qaAbOgaJpZM4JQTLh
.
Regards,
Luiz Vitor Martinez Cardoso
CEO & Founder at geeksys.com.br http://www.geeksys.com.br
Phone: (11) 97351-7097 | Skype: grabberbr
"If you wanna be successful, you need total dedication, go for your last
limit, give your best and love your love infinitely!"
"The only limits are the ones you place upon yourself"
from squeezenet.
Could uou share your solver.txt?
i no use the default solver.
net: "protoueezenet_face.prototxt" snapshot_prefix: "snapshotueezenet_face_1" base_lr: 0.001 display: 1000 test_interval: 5000 snapshot: 5000 test_iter: 100 lr_policy: "step" gamma: 0.1 stepsize: 100000 max_iter: 300000 momentum: 0.9 weight_decay: 0.0002 solver_mode: GPU
from squeezenet.
@bluekingdom thank you! On Darknet framework it is nothing working either... loss at a certain point (5~6) stop decreasing.
from squeezenet.
Same problem, learning very unstable.
I have tested it on 20k images from https://www.kaggle.com/c/dogs-vs-cats
At the start of the training for 5-10 epochs learning curve looks like dead, but than it start learning. Also even worse that I have successfully trained it once but then can't reproduce results with same settings!
from squeezenet.
@mrgloom what framework are you using for training?
from squeezenet.
I'm using NVIDIA DIGITS with Caffe backend.
from squeezenet.
Hey, I used the same data set.
https://www.kaggle.com/c/dogs-vs-cats
and my solver.prototxt is shown below.
test_iter: 1000 test_interval: 1000 base_lr: 0.04 display: 40 max_iter: 100000 iter_size: 16 lr_policy: "poly" power: 1.0 momentum: 0.9 weight_decay: 0.0002 snapshot: 10000 snapshot_prefix: "D:/deeplearning-cats-dogs-tutorial/caffe_models/SqueezeNet/SqueezeNet" solver_mode: GPU random_seed: 42 net: "D:/deeplearning-cats-dogs-tutorial/caffe_models/SqueezeNet/train_val_v1.1.prototxt" test_initialization: false average_loss: 40
Please see attached learning curve.
The accuracy is only <0.8 even though it is just 2-class classification.
Are there any problems in my solver setting?
from squeezenet.
You can check this out
https://github.com/mrgloom/kaggle-dogs-vs-cats-solution/tree/master/learning_from_scratch/Models/SqeezeNet_v1.1
However, I can't reproduce results with same settings.
from squeezenet.
I used solver.prototxt shown below.
test_iter: 100 test_interval: 1000 base_lr: 0.001 display: 100 max_iter: 30000 iter_size: 16 lr_policy: "poly" power: 1.0 momentum: 0.9 weight_decay: 0.0002 snapshot: 1000 snapshot_prefix: "D:/deeplearning-cats-dogs-tutorial/caffe_models/SqueezeNet/SqueezeNet" solver_mode: GPU net: "D:/deeplearning-cats-dogs-tutorial/caffe_models/SqueezeNet/train_val_v1.1.prototxt" test_initialization: false
I am not sure if this is the best for SqueezeNet 2-class classification.
from squeezenet.
For the people who are experimenting with Dogs vs Cats... this person did some experiments with SqueezeNet and other models for a similar challenge:
https://florianbordes.wordpress.com/2016/04/16/cats-vs-dogs-12-summary-and-conclusion/
from squeezenet.
Here is working example
https://github.com/mrgloom/kaggle-dogs-vs-cats-solution/tree/master/learning_from_scratch/Models/SqeezeNet_v1.1
from squeezenet.
Related Issues (20)
- Has anyone successfully trained Squeezenet with residual connections?
- model convert HOT 1
- SqueezeNet v1.1 with Residual Connections with Dense→Sparse→Dense (DSD) Training
- Top-1 Acc=61.0% on ImageNet, without any sacrificing compared with SqueezeNet v1.1. HOT 4
- Image Width Issue HOT 1
- tensorflow- After hundreds of epochs, my total_loss stay around 0.6~0.7, and not decreased HOT 2
- 1.1 deploy.prototxt HOT 1
- SqueezeNet is slower when using GPU than when using CPU? HOT 2
- training from scratch, random seed HOT 1
- why can not get the output of the prob layer? HOT 1
- SqueezeNet training on cifar HOT 3
- The SqueezeNet deploy.caffemodel files have all 0.0 weight and bias data HOT 1
- Fine-tuning SqueezeNet HOT 2
- which label list you used HOT 1
- why not use lr_mult, decay_mult like {1, 1, 2, 0}? HOT 3
- optimization and compression
- Image normalization values HOT 1
- squeezenet v1_1 for facedetector , possible , feasable ? HOT 1
- squeezenet for speech HOT 3
- Some minor mistakes in the paper HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from squeezenet.