Comments (2)
@tom-schoener - yes, that's expected - there are no adapter weights in the original pre-trained bert checkpoints (i.e. those from google-research/bert).
More concerning is however the "trainable params: 0" line in the summary. To fix this, please, put the
l_bert.apply_adapter_freeze()
only once the model has been build (i.e. after the model.build()
, which would instantiate the properly sized weights).
And as a side note - usually adapter_size
does not have to be too big. Depending on the task as small as 4 or 8 could be sufficient, and sometimes even without an adapter, freezing all of bert and tweaking only the layer_normalization layers could work surprisingly well.
from bert-for-tf2.
Thank you for the quick response. The problem with the number of trainable weights being 0 was just a copy-paste error for the code example.
I tried your suggested adapter_size of 4 and it works quite well for my task. Also, freezing all layers and only tweaking the normalization layers sounds interesting. I am going to try that as well.
I really like your BERT implementation for TF Keras - keep up the good work!
from bert-for-tf2.
Related Issues (20)
- Custom tokenizer layer HOT 5
- ResourceExhaustedError: OOM when allocating tensor with shape[501153,768] and type float on /job:localhost/replica:0/task:0/device:GPU:0 by allocator GPU_0_bfc [Op:Mul]
- mixed precision HOT 3
- example (gpu_movie_reviews) has some mistake
- Failed to get weights from pretrained google model HOT 2
- Can not load pretrained bert weights when loading chinese_L-12_H-768_A-12/bert_model.ckpt HOT 3
- Paddings must be non-negative
- albert classification error(Failed copying input tensor from GPU in order to run Identity: GPU sync failed [Op:Identity])
- ValueError: Found unexpected keys that do not correspond to any Model output
- More comments for the code
- Can't train BERT with loaded weights on QA Task HOT 3
- Setting unexpected parameter 'name' in Params instance 'Params' HOT 2
- how to using this in functional model
- may be there is some problem work with tf hub
- AttributeError: module 'bert' has no attribute 'Layer'
- type error HOT 5
- Activation after bert-layer differs
- Count of weight not found[196]
- OSS License compatibility question
- tensorflow.python.keras.layer.input_spec should be replaced with tensorflow.keras.layers.InputSpec HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from bert-for-tf2.