Update model_config.yaml
Browse files- model_config.yaml +2 -16
model_config.yaml
CHANGED
|
@@ -33,7 +33,6 @@ model:
|
|
| 33 |
attention_bias: false
|
| 34 |
attention_dropout: 0.0
|
| 35 |
mlp_bias: false
|
| 36 |
-
|
| 37 |
tokenizer:
|
| 38 |
type: "sentencepiece"
|
| 39 |
model_max_length: 131072
|
|
@@ -42,15 +41,6 @@ model:
|
|
| 42 |
chat_template: "{% for message in messages %}{{ '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>\n' }}{% endfor %}{{ '<|im_start|>assistant\n' }}"
|
| 43 |
|
| 44 |
training:
|
| 45 |
-
base_model: "meta-llama/Meta-Llama-3.1-70B"
|
| 46 |
-
training_data:
|
| 47 |
-
- "scientific_papers"
|
| 48 |
-
- "code_repositories"
|
| 49 |
-
- "mathematical_proofs"
|
| 50 |
-
- "conversational_data"
|
| 51 |
-
- "multilingual_corpus"
|
| 52 |
-
- "technical_documentation"
|
| 53 |
-
total_tokens: "2.5T"
|
| 54 |
training_steps: 150000
|
| 55 |
warmup_steps: 2000
|
| 56 |
learning_rate: 2.0e-5
|
|
@@ -70,14 +60,10 @@ model:
|
|
| 70 |
|
| 71 |
quantization:
|
| 72 |
bits: 16
|
|
|
|
| 73 |
supported_formats:
|
| 74 |
- "fp16"
|
| 75 |
-
|
| 76 |
-
- "int8"
|
| 77 |
-
- "int4"
|
| 78 |
-
- "awq"
|
| 79 |
-
- "gptq"
|
| 80 |
-
- "gguf"
|
| 81 |
|
| 82 |
inference:
|
| 83 |
default_parameters:
|
|
|
|
| 33 |
attention_bias: false
|
| 34 |
attention_dropout: 0.0
|
| 35 |
mlp_bias: false
|
|
|
|
| 36 |
tokenizer:
|
| 37 |
type: "sentencepiece"
|
| 38 |
model_max_length: 131072
|
|
|
|
| 41 |
chat_template: "{% for message in messages %}{{ '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>\n' }}{% endfor %}{{ '<|im_start|>assistant\n' }}"
|
| 42 |
|
| 43 |
training:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
training_steps: 150000
|
| 45 |
warmup_steps: 2000
|
| 46 |
learning_rate: 2.0e-5
|
|
|
|
| 60 |
|
| 61 |
quantization:
|
| 62 |
bits: 16
|
| 63 |
+
precision: "float16"
|
| 64 |
supported_formats:
|
| 65 |
- "fp16"
|
| 66 |
+
note: "Model is provided in full FP16 precision without quantization"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
inference:
|
| 69 |
default_parameters:
|