Whisper Large-v3 β Fine-tuned on Slovak Plenary ASR Corpus
This model is a fine-tuned version of openai/whisper-large-v3.
It is adapted for Slovak ASR using SloPalSpeech: 2,806 hours of aligned, β€30 s speechβtext pairs from official plenary sessions of the Slovak National Council.
- Language: Slovak
 - Domain: Parliamentary / formal speech
 - Training data: 2,806 h
 - Intended use: Slovak speech recognition; strongest in formal/public-speaking contexts
 
π§ͺ Evaluation
| Dataset | Base WER | Fine-tuned WER | Ξ (abs) | 
|---|---|---|---|
| Common Voice 21 (sk) | 20.8 | 11.6 | -9.2 | 
| FLEURS (sk) | 9.2 | 5.5 | -3.7 | 
Numbers from the paperβs final benchmark runs.
π§ Training Details
- Framework: Hugging Face Transformers
 - Hardware: Multi-GPU setup (NVIDIA A10s) with Fully Sharded Data Parallel (FSDP)
 - Epochs: ~2 with early stopping on validation WER
 - Learning rate: 
1e-5with weight decay0.01to prevent overfitting - Notes: Training required sharded checkpoints; evaluation run separately due to runtime compatibility issues
 
β οΈ Limitations
- Domain bias toward parliamentary speech (e.g., political vocabulary, formal register).
 - As with Whisper models generally, occasional hallucinations may appear; consider temperature fallback / compression-ratio checks at inference time.
 - Multilingual performance is not guaranteed (full-parameter finetuning emphasized Slovak).
 
π Citation & Paper
For more details, please see our paper on arXiv. If you use this model in your work, please cite it as:
@misc{boΕΎΓk2025slopalspeech2800hourslovakspeech,
      title={SloPalSpeech: A 2,800-Hour Slovak Speech Corpus from Parliamentary Data}, 
      author={Erik BoΕΎΓk and Marek Ε uppa},
      year={2025},
      eprint={2509.19270},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2509.19270}, 
}
π Acknowledgements
This work was supported by VΓB Banka who provided the GPU resources and backing necessary to accomplish it, enabling progress in Slovak ASR research.
- Downloads last month
 - 49
 
	Inference Providers
	NEW
	
	
	This model isn't deployed by any Inference Provider.
	π
			
		Ask for provider support
Model tree for erikbozik/whisper-large-v3-sk
Base model
openai/whisper-large-v3Dataset used to train erikbozik/whisper-large-v3-sk
Evaluation results
- WER on Common Voice 21 (Slovak test set)self-reported11.600
 - WER on FLEURS (Slovak test set)self-reported5.500