Comments (7)
Hi Stefan,
Great to see non-English models being trained already! Here are some pre-training curves (x-axis is pre-train steps) for the models. Note that these metrics were computed with "do_eval": true
which means there is no dropout, so training losses will be slightly higher.
Small model on OpenWebText, batch_size=128, mask_percent=0.15
Base model on WikiBooks, batch_size=256, mask_percent=0.15
Large model on XLNet data, batch_size=2048, mask_percent=0.25
The losses will vary depending on your setting because the disciminator loss depends on the quality of the generator, the generator loss can depend on the training data., etc.
from electra.
Hi,
Great to have these training curves! I'm currently working on a new non-English ELECTRA large model. I encountered some problems during pre-training. The loss first drops after a period of time and then rises. I would just like to ask if you can post the learning rate of both base and large models (or even share the learning rate curve) so that we have a kind of reference point about learning rate when training own models . The following is my loss and acc curve. Looking forward to your suggestions.
Thanks many in advance!
from electra.
@clarkkev Do you have any idea about why there's a peak in disc_loss at the very beginning of the training curve?
from electra.
@yaolu From our own experiments, the generator starts out creating random predictions, then plateaus for a bit using the median prediction. So every [MASK] gets replaced with the token the
or the .
token. This makes the discriminator's job very easy for a bit. Eventually the generator moves past this plateau and starts generating more diverse predictions, which causes the discriminator loss to peak for a bit.
Here's a sample validation prediction at step 1000. The red-highlighted tokens are the generator's replacements, and the yellow-highlighted tokens are the tokens which the discriminator predicts as corrupted. Note that the discriminator simply predicts that every the
and .
token is corrupt, and the generator in this case replaces every [MASK]
with a .
.
from electra.
Hi @clarkkev,
thanks for these interesting insights! Great to have these training curves 🤗
from electra.
Hi,
Great to have these training curves! I'm currently working on a new non-English ELECTRA large model. I encountered some problems during pre-training. The loss first drops after a period of time and then rises. I would just like to ask if you can post the learning rate of both base and large models (or even share the learning rate curve) so that we have a kind of reference point about learning rate when training own models . The following is my loss and acc curve. Looking forward to your suggestions.Thanks many in advance!
How did you draw the graph?
When I set the do_eval as true and evaluated the model, it was marked as only one point when I checked with Tensorboard.
How can I check the graph every step?
from electra.
as only one poin
I also can not draw this graph, are you solve it?
from electra.
Related Issues (20)
- what should i do to extract the electra discriminator HOT 2
- ELECTRA-base fine tuned on MNLI HOT 1
- A possible mistake in the FLOPs calculation of attn_output_layer_norm in the file flops_computation.py
- no module named tensorflow.contrib HOT 1
- Electra Vocabulary HOT 1
- some confusions about paper HOT 1
- Question regarding TrainingsData/Validation Data split
- NumPy Import Error HOT 2
- Train electra with another tokenizer HOT 2
- About the Electra paper
- Optimal Learning Rate and Training Steps for Large Batch Size
- How can I draw this? HOT 1
- Tagging Task Segment ids
- Cannot import trace from tensorflow.python.profiler HOT 4
- sequence tagging tasks fails at metric reporting HOT 1
- Can I used run_mlm.py in transformer for fine-tuning generator(mlm) of electra
- failed to run cuBLAS routine: CUBLAS_STATUS_EXECUTION_FAILED HOT 1
- finetune preprocessing adding padding to the dataset error HOT 1
- How many parameters do discriminator and generator have?
- What is the maximum acceptance for the sentence length for the ELECTRA model?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from electra.