Add metadata and links to model card (#16)
Browse files- Add metadata and links to model card (57d6888337c9e6e7bbcffa3237fb5b33c6c7227a)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
|
@@ -1,31 +1,32 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: mit
|
| 3 |
license_link: https://huggingface.co/microsoft/Phi-4-mini-instruct/resolve/main/LICENSE
|
| 4 |
-
language:
|
| 5 |
-
- "multilingual"
|
| 6 |
-
- "ar"
|
| 7 |
-
- "zh"
|
| 8 |
-
- "cs"
|
| 9 |
-
- "da"
|
| 10 |
-
- "nl"
|
| 11 |
-
- "en"
|
| 12 |
-
- "fi"
|
| 13 |
-
- "fr"
|
| 14 |
-
- "de"
|
| 15 |
-
- "he"
|
| 16 |
-
- "hu"
|
| 17 |
-
- "it"
|
| 18 |
-
- "ja"
|
| 19 |
-
- "ko"
|
| 20 |
-
- "no"
|
| 21 |
-
- "pl"
|
| 22 |
-
- "pt"
|
| 23 |
-
- "ru"
|
| 24 |
-
- "es"
|
| 25 |
-
- "sv"
|
| 26 |
-
- "th"
|
| 27 |
-
- "tr"
|
| 28 |
-
- "uk"
|
| 29 |
pipeline_tag: text-generation
|
| 30 |
tags:
|
| 31 |
- nlp
|
|
@@ -34,9 +35,8 @@ widget:
|
|
| 34 |
- messages:
|
| 35 |
- role: user
|
| 36 |
content: Can you provide ways to eat combinations of bananas and dragonfruits?
|
| 37 |
-
library_name: transformers
|
| 38 |
---
|
| 39 |
-
|
| 40 |
## Model Summary
|
| 41 |
|
| 42 |
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
|
|
@@ -46,6 +46,8 @@ Phi-4-mini-instruct is a lightweight open model built upon synthetic data and fi
|
|
| 46 |
π©βπ³ [Phi Cookbook](https://github.com/microsoft/PhiCookBook) <br>
|
| 47 |
π‘ [Phi Portal](https://azure.microsoft.com/en-us/products/phi) <br>
|
| 48 |
π₯οΈ Try It [Azure](https://aka.ms/phi-4-mini/azure), [Huggingface](https://huggingface.co/spaces/microsoft/phi-4-mini) <br>
|
|
|
|
|
|
|
| 49 |
|
| 50 |
π**Phi-4**: [[multimodal-instruct](https://huggingface.co/microsoft/Phi-4-multimodal-instruct) | [onnx](https://huggingface.co/microsoft/Phi-4-multimodal-instruct-onnx)];
|
| 51 |
[[mini-instruct](https://huggingface.co/microsoft/Phi-4-mini-instruct) | [onnx](https://huggingface.co/microsoft/Phi-4-mini-instruct-onnx)]
|
|
@@ -368,4 +370,4 @@ The model was evaluated across a breadth of public and internal benchmarks to un
|
|
| 368 |
+ XSTest: exaggerated safety evaluation
|
| 369 |
+ Toxigen: adversarial and hate speech detection
|
| 370 |
+ Red Team:
|
| 371 |
-
+ Responses to prompts provided by AI Red Team at Microsoft
|
|
|
|
| 1 |
---
|
| 2 |
+
language:
|
| 3 |
+
- multilingual
|
| 4 |
+
- ar
|
| 5 |
+
- zh
|
| 6 |
+
- cs
|
| 7 |
+
- da
|
| 8 |
+
- nl
|
| 9 |
+
- en
|
| 10 |
+
- fi
|
| 11 |
+
- fr
|
| 12 |
+
- de
|
| 13 |
+
- he
|
| 14 |
+
- hu
|
| 15 |
+
- it
|
| 16 |
+
- ja
|
| 17 |
+
- ko
|
| 18 |
+
- 'no'
|
| 19 |
+
- pl
|
| 20 |
+
- pt
|
| 21 |
+
- ru
|
| 22 |
+
- es
|
| 23 |
+
- sv
|
| 24 |
+
- th
|
| 25 |
+
- tr
|
| 26 |
+
- uk
|
| 27 |
+
library_name: transformers
|
| 28 |
license: mit
|
| 29 |
license_link: https://huggingface.co/microsoft/Phi-4-mini-instruct/resolve/main/LICENSE
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
pipeline_tag: text-generation
|
| 31 |
tags:
|
| 32 |
- nlp
|
|
|
|
| 35 |
- messages:
|
| 36 |
- role: user
|
| 37 |
content: Can you provide ways to eat combinations of bananas and dragonfruits?
|
|
|
|
| 38 |
---
|
| 39 |
+
|
| 40 |
## Model Summary
|
| 41 |
|
| 42 |
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
|
|
|
|
| 46 |
π©βπ³ [Phi Cookbook](https://github.com/microsoft/PhiCookBook) <br>
|
| 47 |
π‘ [Phi Portal](https://azure.microsoft.com/en-us/products/phi) <br>
|
| 48 |
π₯οΈ Try It [Azure](https://aka.ms/phi-4-mini/azure), [Huggingface](https://huggingface.co/spaces/microsoft/phi-4-mini) <br>
|
| 49 |
+
|
| 50 |
+
π [Model paper](https://huggingface.co/papers/2503.01743)
|
| 51 |
|
| 52 |
π**Phi-4**: [[multimodal-instruct](https://huggingface.co/microsoft/Phi-4-multimodal-instruct) | [onnx](https://huggingface.co/microsoft/Phi-4-multimodal-instruct-onnx)];
|
| 53 |
[[mini-instruct](https://huggingface.co/microsoft/Phi-4-mini-instruct) | [onnx](https://huggingface.co/microsoft/Phi-4-mini-instruct-onnx)]
|
|
|
|
| 370 |
+ XSTest: exaggerated safety evaluation
|
| 371 |
+ Toxigen: adversarial and hate speech detection
|
| 372 |
+ Red Team:
|
| 373 |
+
+ Responses to prompts provided by AI Red Team at Microsoft
|