Update README.md
Browse files
README.md
CHANGED
|
@@ -49,6 +49,20 @@ This model was trained with DPO, a method introduced in [Direct Preference Optim
|
|
| 49 |
|
| 50 |
## Citations
|
| 51 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 52 |
Cite DPO as:
|
| 53 |
|
| 54 |
```bibtex
|
|
|
|
| 49 |
|
| 50 |
## Citations
|
| 51 |
|
| 52 |
+
Cite this model:
|
| 53 |
+
|
| 54 |
+
```bibtex
|
| 55 |
+
@misc{lübbers2025enhancingparaphrasetypegeneration,
|
| 56 |
+
title={Enhancing Paraphrase Type Generation: The Impact of DPO and RLHF Evaluated with Human-Ranked Data},
|
| 57 |
+
author={Christopher Lee Lübbers},
|
| 58 |
+
year={2025},
|
| 59 |
+
eprint={2506.02018},
|
| 60 |
+
archivePrefix={arXiv},
|
| 61 |
+
primaryClass={cs.CL},
|
| 62 |
+
url={https://arxiv.org/abs/2506.02018},
|
| 63 |
+
}
|
| 64 |
+
```
|
| 65 |
+
|
| 66 |
Cite DPO as:
|
| 67 |
|
| 68 |
```bibtex
|