Update config.json
Browse files- config.json +8 -20
config.json
CHANGED
|
@@ -1,29 +1,17 @@
|
|
| 1 |
{
|
| 2 |
-
"model_name": "slim-
|
| 3 |
-
"model_ft_base": "slim-
|
| 4 |
"quantization": "4Q_K_M GGUF",
|
| 5 |
"model_base": "stabilityai/stablelm-3b-4e1t",
|
| 6 |
"model_type": "stablelm",
|
| 7 |
-
"tokenizer": "llmware/slim-
|
| 8 |
"parameters": "2.7 billion",
|
| 9 |
-
"description": "slim-
|
| 10 |
"prompt_wrapper": "human_bot",
|
| 11 |
-
"prompt_format": "<human> {context_passage} <classify>
|
| 12 |
-
"output_format": "{
|
| 13 |
-
"primary_keys": [
|
| 14 |
-
|
| 15 |
-
"people",
|
| 16 |
-
"place",
|
| 17 |
-
"organization",
|
| 18 |
-
"misc"
|
| 19 |
-
],
|
| 20 |
-
"output_values": [
|
| 21 |
-
"sentiment",
|
| 22 |
-
"people",
|
| 23 |
-
"place",
|
| 24 |
-
"organization",
|
| 25 |
-
"misc"
|
| 26 |
-
],
|
| 27 |
"publisher": "llmware",
|
| 28 |
"release_date": "march 2024",
|
| 29 |
"test_set": [
|
|
|
|
| 1 |
{
|
| 2 |
+
"model_name": "slim-extract-tool",
|
| 3 |
+
"model_ft_base": "slim-extract",
|
| 4 |
"quantization": "4Q_K_M GGUF",
|
| 5 |
"model_base": "stabilityai/stablelm-3b-4e1t",
|
| 6 |
"model_type": "stablelm",
|
| 7 |
+
"tokenizer": "llmware/slim-extract",
|
| 8 |
"parameters": "2.7 billion",
|
| 9 |
+
"description": "slim-extract is a function-calling model, fine-tuned to output structured dictionaries corresponding to a custom extraction key",
|
| 10 |
"prompt_wrapper": "human_bot",
|
| 11 |
+
"prompt_format": "<human> {context_passage} <classify> {custom extraction key} </classify>\n<bot>:",
|
| 12 |
+
"output_format": "{custom_extraction_key: ['list of items found in the text corresponding to key']}",
|
| 13 |
+
"primary_keys": ["any custom key provided"],
|
| 14 |
+
"output_values": ["dictionary with custom_key and value consisting of list of extracted values"],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
"publisher": "llmware",
|
| 16 |
"release_date": "march 2024",
|
| 17 |
"test_set": [
|