dbln / stochastic_depth_keras Goto Github PK
View Code? Open in Web Editor NEWKeras implementation for "Deep Networks with Stochastic Depth" http://arxiv.org/abs/1603.09382
License: MIT License
Keras implementation for "Deep Networks with Stochastic Depth" http://arxiv.org/abs/1603.09382
License: MIT License
Thanks a lot for doing this. I might misunderstand something, but the chart seems to say 15% eventual validation error on CIFAR-10. The original yueatsprograms Torch implementation has 5.23% validation error. I believe this difference is too large to be attributed to the different number of resnet blocks, or the lack of augmentation. Is there some missing functionality or known bug responsible for the discrepancy?
Thanks a lot for doing this. I might misunderstand something, but l can understand how the gate update when training. I can not understand the two code...
gate = K.variable(1, dtype="uint8")
add_tables += [{"death_rate": _death_rate, "gate": gate}]
return Lambda(lambda tensors: K.switch(gate, tensors[0], tensors[1]),
output_shape=output_shape)([out, x])
Is this 'gate' always equal 1,when training ?...
class GatesUpdate(Callback):
def on_batch_begin(self, batch, logs={}):
open_all_gates()
rands = np.random.uniform(size=len(add_tables))
for t, rand in zip(add_tables, rands):
if rand < K.get_value(t["death_rate"]):
K.set_value(t["gate"], 0)
Is this 'GatesUpdate' make action on the 'Lambda' layer ,when training?
Thank you.
https://github.com/dblN/stochastic_depth_keras/blob/d26c492/train.py#L86-L87
When doing lin_decay, these two lines set the scaling to one minus the maximal value of the death rate. I believe they should be set to one minus the death rate of the current residual block. Am I misuderstanding something?
Hi,
Following your advice regarding setting the recursion limit.
I managed to get this to run with N = 17. My windows machine has a 16GB of RAM, using the Theano backend. higher than that, python crashes.
Here are my questions:
Thank you!
Hi, I got a RuntimeError: maximum recursion depth exceeded in cmp
when running it on a virtualenv
. It seems related to Theano Issue #689.
I have the following libraries installed on my virtual environment (Keras is installed from keras-1
branch):
Keras==1.0.0
numpy==1.11.0
PyYAML==3.11
scipy==0.17.0
six==1.10.0
Theano==0.8.1
You are using image tensors with the Theano dimension ordering conventions: (samples, channels, width, height)
. You want to do BatchNorm on the channels, therefore you should use:
BatchNormalization(axis=1)
instead of:
BatchNormalization() # default for axis is -1
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.