Difference between revisions of "Eosc/NorBERT3 corpus"
(Cleaning) |
(→To Decide) |
||
Line 13: | Line 13: | ||
== To Decide == | == To Decide == | ||
− | + | The size of NBDigital is 662M tokens. Should we use it? It probably overlaps a lot with NCC. |
Revision as of 21:41, 18 October 2022
Workflow
Sampling experiment
We plan to create two versions of the training corpus:
- baseline (as is)
- Wikipedia+NCC+NAK multiplied by two to match the C4 size (oversampling quality data)
Vocabulary
Starting with 50K, following NorBERT-2. May be later experiment with other values.
To Decide
The size of NBDigital is 662M tokens. Should we use it? It probably overlaps a lot with NCC.