Update README.md
Browse files
README.md
CHANGED
|
@@ -54,3 +54,18 @@ speaker_embedding = model(dummy_audio)
|
|
| 54 |
print("\nEmbedding extracted successfully!")
|
| 55 |
print(f"Output embedding shape: {speaker_embedding.shape}")
|
| 56 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
print("\nEmbedding extracted successfully!")
|
| 55 |
print(f"Output embedding shape: {speaker_embedding.shape}")
|
| 56 |
```
|
| 57 |
+
## Citation
|
| 58 |
+
|
| 59 |
+
If you find MHFA useful, please cite it as
|
| 60 |
+
|
| 61 |
+
```bibtex
|
| 62 |
+
@inproceedings{peng2023attention,
|
| 63 |
+
title={An attention-based backend allowing efficient fine-tuning of transformer models for speaker verification},
|
| 64 |
+
author={Peng, Junyi and Plchot, Old{\v{r}}ich and Stafylakis, Themos and Mo{\v{s}}ner, Ladislav and Burget, Luk{\'a}{\v{s}} and {\v{C}}ernock{\`y}, Jan},
|
| 65 |
+
booktitle={2022 IEEE Spoken Language Technology Workshop (SLT)},
|
| 66 |
+
pages={555--562},
|
| 67 |
+
year={2023},
|
| 68 |
+
organization={IEEE}
|
| 69 |
+
}
|
| 70 |
+
|
| 71 |
+
```
|