Update README.md (#1)
Browse files- Update README.md (df23ef4cb7ab63453bc89bcc53a6e634d91664b9)
Co-authored-by: Elodie Gauthier <gauthelo@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -4,7 +4,7 @@ license: cc-by-nc-4.0
|
|
| 4 |
|
| 5 |
|
| 6 |
## Model description
|
| 7 |
-
This self-supervised speech model (a.k.a. SSA-HuBERT-
|
| 8 |
It was trained on nearly 60 000 hours of speech segments and covers 21 languages and variants spoken in Sub-Saharan Africa.
|
| 9 |
|
| 10 |
### Pretraining data
|
|
@@ -15,7 +15,7 @@ It was trained on nearly 60 000 hours of speech segments and covers 21 languages
|
|
| 15 |
## ASR fine-tuning
|
| 16 |
The SpeechBrain toolkit (Ravanelli et al., 2021) is used to fine-tune the model.
|
| 17 |
Fine-tuning is done for each language using the FLEURS dataset [2].
|
| 18 |
-
The pretrained model (SSA-HuBERT-
|
| 19 |
|
| 20 |
## License
|
| 21 |
This model is released under the CC-by-NC 4.0 conditions.
|
|
|
|
| 4 |
|
| 5 |
|
| 6 |
## Model description
|
| 7 |
+
This self-supervised speech model (a.k.a. SSA-HuBERT-XL-60k) is based on a HuBERT X-Large architecture (~964M params) [1].
|
| 8 |
It was trained on nearly 60 000 hours of speech segments and covers 21 languages and variants spoken in Sub-Saharan Africa.
|
| 9 |
|
| 10 |
### Pretraining data
|
|
|
|
| 15 |
## ASR fine-tuning
|
| 16 |
The SpeechBrain toolkit (Ravanelli et al., 2021) is used to fine-tune the model.
|
| 17 |
Fine-tuning is done for each language using the FLEURS dataset [2].
|
| 18 |
+
The pretrained model (SSA-HuBERT-XL-60k) is considered as a speech encoder and is fully fine-tuned with two 1024 linear layers and a softmax output at the top.
|
| 19 |
|
| 20 |
## License
|
| 21 |
This model is released under the CC-by-NC 4.0 conditions.
|