Update config.json
Browse files- config.json +0 -2
config.json
CHANGED
|
@@ -1,5 +1,4 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "/data_storage/lbw/MMaDA/mmada-training-stage3-llada-instruct-512-cot-uni/checkpoint-210000/unwrapped_model",
|
| 3 |
"activation_type": "silu",
|
| 4 |
"alibi": false,
|
| 5 |
"alibi_bias_max": 8.0,
|
|
@@ -47,7 +46,6 @@
|
|
| 47 |
"num_vq_tokens": 256,
|
| 48 |
"pad_token_id": 126081,
|
| 49 |
"precision": "amp_bf16",
|
| 50 |
-
"pretrained_model_path": "/data_storage/shared/pretrained_models/LLaDA-8B-Instruct",
|
| 51 |
"residual_dropout": 0.0,
|
| 52 |
"rms_norm_eps": 1e-05,
|
| 53 |
"rope": true,
|
|
|
|
| 1 |
{
|
|
|
|
| 2 |
"activation_type": "silu",
|
| 3 |
"alibi": false,
|
| 4 |
"alibi_bias_max": 8.0,
|
|
|
|
| 46 |
"num_vq_tokens": 256,
|
| 47 |
"pad_token_id": 126081,
|
| 48 |
"precision": "amp_bf16",
|
|
|
|
| 49 |
"residual_dropout": 0.0,
|
| 50 |
"rms_norm_eps": 1e-05,
|
| 51 |
"rope": true,
|