Update README.md
Browse files
README.md
CHANGED
|
@@ -7,6 +7,7 @@ library_name: peft
|
|
| 7 |
co2_eq_emissions: 110
|
| 8 |
base_model:
|
| 9 |
- mistralai/Mistral-7B-v0.1
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
# Adapter for Mistral-7B-v0.1 fine-tuned on DFP
|
|
@@ -254,16 +255,18 @@ trainer.train()
|
|
| 254 |
|
| 255 |
## Citations
|
| 256 |
### PEFT library
|
| 257 |
-
|
|
|
|
| 258 |
title = {PEFT: State-of-the-art Parameter-Efficient Fine-Tuning methods},
|
| 259 |
author = {Sourab Mangrulkar and Sylvain Gugger and Lysandre Debut and Younes Belkada and Sayak Paul and Benjamin Bossan},
|
| 260 |
howpublished = {\url{https://github.com/huggingface/peft}},
|
| 261 |
year = {2022}
|
| 262 |
}
|
|
|
|
| 263 |
|
| 264 |
### Mistral-7B-Instruct-v0.1
|
| 265 |
-
|
| 266 |
-
|
| 267 |
title={Mistral 7B},
|
| 268 |
author={Albert Q. Jiang and Alexandre Sablayrolles and Arthur Mensch and Chris Bamford and Devendra Singh Chaplot and Diego de las Casas and Florian Bressand and Gianna Lengyel and Guillaume Lample and Lucile Saulnier and Lélio Renard Lavaud and Marie-Anne Lachaux and Pierre Stock and Teven Le Scao and Thibaut Lavril and Thomas Wang and Timothée Lacroix and William El Sayed},
|
| 269 |
year={2023},
|
|
@@ -271,9 +274,11 @@ trainer.train()
|
|
| 271 |
archivePrefix={arXiv},
|
| 272 |
primaryClass={cs.CL}
|
| 273 |
}
|
|
|
|
| 274 |
|
| 275 |
### DFP
|
| 276 |
-
|
|
|
|
| 277 |
author = { {Centre Aquitain des Technologies de l'Information et Electroniques} },
|
| 278 |
title = { DFP (Revision 1d24c09) },
|
| 279 |
year = 2023,
|
|
@@ -281,14 +286,16 @@ trainer.train()
|
|
| 281 |
doi = { 10.57967/hf/1200 },
|
| 282 |
publisher = { Hugging Face }
|
| 283 |
}
|
| 284 |
-
|
| 285 |
|
| 286 |
### LoRA
|
| 287 |
-
|
|
|
|
| 288 |
title={LoRA: Low-Rank Adaptation of Large Language Models},
|
| 289 |
author={Edward J. Hu and Yelong Shen and Phillip Wallis and Zeyuan Allen-Zhu and Yuanzhi Li and Shean Wang and Lu Wang and Weizhu Chen},
|
| 290 |
year={2021},
|
| 291 |
eprint={2106.09685},
|
| 292 |
archivePrefix={arXiv},
|
| 293 |
primaryClass={cs.CL}
|
| 294 |
-
}
|
|
|
|
|
|
| 7 |
co2_eq_emissions: 110
|
| 8 |
base_model:
|
| 9 |
- mistralai/Mistral-7B-v0.1
|
| 10 |
+
pipeline_tag: text-generation
|
| 11 |
---
|
| 12 |
|
| 13 |
# Adapter for Mistral-7B-v0.1 fine-tuned on DFP
|
|
|
|
| 255 |
|
| 256 |
## Citations
|
| 257 |
### PEFT library
|
| 258 |
+
```
|
| 259 |
+
@Misc{peft,
|
| 260 |
title = {PEFT: State-of-the-art Parameter-Efficient Fine-Tuning methods},
|
| 261 |
author = {Sourab Mangrulkar and Sylvain Gugger and Lysandre Debut and Younes Belkada and Sayak Paul and Benjamin Bossan},
|
| 262 |
howpublished = {\url{https://github.com/huggingface/peft}},
|
| 263 |
year = {2022}
|
| 264 |
}
|
| 265 |
+
```
|
| 266 |
|
| 267 |
### Mistral-7B-Instruct-v0.1
|
| 268 |
+
```
|
| 269 |
+
@misc{jiang2023mistral,
|
| 270 |
title={Mistral 7B},
|
| 271 |
author={Albert Q. Jiang and Alexandre Sablayrolles and Arthur Mensch and Chris Bamford and Devendra Singh Chaplot and Diego de las Casas and Florian Bressand and Gianna Lengyel and Guillaume Lample and Lucile Saulnier and Lélio Renard Lavaud and Marie-Anne Lachaux and Pierre Stock and Teven Le Scao and Thibaut Lavril and Thomas Wang and Timothée Lacroix and William El Sayed},
|
| 272 |
year={2023},
|
|
|
|
| 274 |
archivePrefix={arXiv},
|
| 275 |
primaryClass={cs.CL}
|
| 276 |
}
|
| 277 |
+
```
|
| 278 |
|
| 279 |
### DFP
|
| 280 |
+
```
|
| 281 |
+
@misc {centre_aquitain_des_technologies_de_l'information_et_electroniques_2023,
|
| 282 |
author = { {Centre Aquitain des Technologies de l'Information et Electroniques} },
|
| 283 |
title = { DFP (Revision 1d24c09) },
|
| 284 |
year = 2023,
|
|
|
|
| 286 |
doi = { 10.57967/hf/1200 },
|
| 287 |
publisher = { Hugging Face }
|
| 288 |
}
|
| 289 |
+
```
|
| 290 |
|
| 291 |
### LoRA
|
| 292 |
+
```
|
| 293 |
+
@misc{hu2021lora,
|
| 294 |
title={LoRA: Low-Rank Adaptation of Large Language Models},
|
| 295 |
author={Edward J. Hu and Yelong Shen and Phillip Wallis and Zeyuan Allen-Zhu and Yuanzhi Li and Shean Wang and Lu Wang and Weizhu Chen},
|
| 296 |
year={2021},
|
| 297 |
eprint={2106.09685},
|
| 298 |
archivePrefix={arXiv},
|
| 299 |
primaryClass={cs.CL}
|
| 300 |
+
}
|
| 301 |
+
```
|