aihpi commited on
Commit
37c43cc
·
verified ·
1 Parent(s): 9003a09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -67,7 +67,7 @@ The model is particularly well-suited for applications requiring detailed transc
67
 
68
  WER is computed on LangAge and ESLO, containing interjections, hesitations. This dataset is challenging for ASR models, because it contains many recordings of older people with different voice quality, and because features of spontaneous speech are difficult to transcribe.
69
 
70
- Lower WER indicates better performance.
71
 
72
  | Dataset Category | Sample Size | Whisper Large V3 | FrWhisper | Improvement |
73
  |-----------------|-------------|------------------|----------------|-------------|
@@ -82,7 +82,6 @@ Lower WER indicates better performance.
82
  ### Key Performance Highlights
83
 
84
  - **14.18 percentage points** overall WER improvement compared to Whisper Large V3
85
- - **Median WER for LangAge data**: Improved from 100% to 44.44%
86
  - **Consistent performance** across both training and test data (no overfitting)
87
 
88
  ## Training Details
 
67
 
68
  WER is computed on LangAge and ESLO, containing interjections, hesitations. This dataset is challenging for ASR models, because it contains many recordings of older people with different voice quality, and because features of spontaneous speech are difficult to transcribe.
69
 
70
+ Lower WER indicates better performance. For a fair comparison between Whisper Large V3 and FrWhisper, punctuation and capitalization were ignored for computing WER.
71
 
72
  | Dataset Category | Sample Size | Whisper Large V3 | FrWhisper | Improvement |
73
  |-----------------|-------------|------------------|----------------|-------------|
 
82
  ### Key Performance Highlights
83
 
84
  - **14.18 percentage points** overall WER improvement compared to Whisper Large V3
 
85
  - **Consistent performance** across both training and test data (no overfitting)
86
 
87
  ## Training Details