Antoine-caubriere gauthelo commited on
Commit
812c7f4
·
verified ·
1 Parent(s): 44c535b

Update README.md (#1)

Browse files

- Update README.md (df23ef4cb7ab63453bc89bcc53a6e634d91664b9)


Co-authored-by: Elodie Gauthier <gauthelo@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -4,7 +4,7 @@ license: cc-by-nc-4.0
4
 
5
 
6
  ## Model description
7
- This self-supervised speech model (a.k.a. SSA-HuBERT-base-60k-V2) is based on a HuBERT Base architecture (~95M params) [1].
8
  It was trained on nearly 60 000 hours of speech segments and covers 21 languages and variants spoken in Sub-Saharan Africa.
9
 
10
  ### Pretraining data
@@ -15,7 +15,7 @@ It was trained on nearly 60 000 hours of speech segments and covers 21 languages
15
  ## ASR fine-tuning
16
  The SpeechBrain toolkit (Ravanelli et al., 2021) is used to fine-tune the model.
17
  Fine-tuning is done for each language using the FLEURS dataset [2].
18
- The pretrained model (SSA-HuBERT-base-60k) is considered as a speech encoder and is fully fine-tuned with two 1024 linear layers and a softmax output at the top.
19
 
20
  ## License
21
  This model is released under the CC-by-NC 4.0 conditions.
 
4
 
5
 
6
  ## Model description
7
+ This self-supervised speech model (a.k.a. SSA-HuBERT-XL-60k) is based on a HuBERT X-Large architecture (~964M params) [1].
8
  It was trained on nearly 60 000 hours of speech segments and covers 21 languages and variants spoken in Sub-Saharan Africa.
9
 
10
  ### Pretraining data
 
15
  ## ASR fine-tuning
16
  The SpeechBrain toolkit (Ravanelli et al., 2021) is used to fine-tune the model.
17
  Fine-tuning is done for each language using the FLEURS dataset [2].
18
+ The pretrained model (SSA-HuBERT-XL-60k) is considered as a speech encoder and is fully fine-tuned with two 1024 linear layers and a softmax output at the top.
19
 
20
  ## License
21
  This model is released under the CC-by-NC 4.0 conditions.