End of training
Browse files- README.md +2 -2
- model.safetensors +1 -1
- runs/Jan10_02-23-23_a143c71d380e/events.out.tfevents.1736476015.a143c71d380e.836.0 +3 -0
- special_tokens_map.json +1 -1
- tokenizer.json +2 -2
- tokenizer_config.json +1 -1
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the None dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
@@ -47,7 +47,7 @@ The following hyperparameters were used during training:
|
|
| 47 |
|
| 48 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 49 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 50 |
-
| 0.
|
| 51 |
|
| 52 |
|
| 53 |
### Framework versions
|
|
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the None dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
+
- Loss: 0.4901
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
|
|
| 47 |
|
| 48 |
| Training Loss | Epoch | Step | Validation Loss |
|
| 49 |
|:-------------:|:-----:|:----:|:---------------:|
|
| 50 |
+
| 0.6933 | 1.0 | 88 | 0.4901 |
|
| 51 |
|
| 52 |
|
| 53 |
### Framework versions
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1625422896
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c6af87e4fb2a1d7a15dcc3487a3e2d506d835be8d3f24efb5368dfcbb46a259
|
| 3 |
size 1625422896
|
runs/Jan10_02-23-23_a143c71d380e/events.out.tfevents.1736476015.a143c71d380e.836.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6170afb28dd4ceec7b87b78b35702137d7f533570c624c4676b9fd733c7297dc
|
| 3 |
+
size 8468
|
special_tokens_map.json
CHANGED
|
@@ -9,7 +9,7 @@
|
|
| 9 |
"rstrip": false,
|
| 10 |
"single_word": false
|
| 11 |
},
|
| 12 |
-
"pad_token": "
|
| 13 |
"sep_token": "</s>",
|
| 14 |
"unk_token": "<unk>"
|
| 15 |
}
|
|
|
|
| 9 |
"rstrip": false,
|
| 10 |
"single_word": false
|
| 11 |
},
|
| 12 |
+
"pad_token": "<pad>",
|
| 13 |
"sep_token": "</s>",
|
| 14 |
"unk_token": "<unk>"
|
| 15 |
}
|
tokenizer.json
CHANGED
|
@@ -12,9 +12,9 @@
|
|
| 12 |
},
|
| 13 |
"direction": "Right",
|
| 14 |
"pad_to_multiple_of": null,
|
| 15 |
-
"pad_id":
|
| 16 |
"pad_type_id": 0,
|
| 17 |
-
"pad_token": "
|
| 18 |
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
|
|
|
| 12 |
},
|
| 13 |
"direction": "Right",
|
| 14 |
"pad_to_multiple_of": null,
|
| 15 |
+
"pad_id": 1,
|
| 16 |
"pad_type_id": 0,
|
| 17 |
+
"pad_token": "<pad>"
|
| 18 |
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
tokenizer_config.json
CHANGED
|
@@ -50,7 +50,7 @@
|
|
| 50 |
"extra_special_tokens": {},
|
| 51 |
"mask_token": "<mask>",
|
| 52 |
"model_max_length": 1000000000000000019884624838656,
|
| 53 |
-
"pad_token": "
|
| 54 |
"sep_token": "</s>",
|
| 55 |
"tokenizer_class": "BartTokenizer",
|
| 56 |
"trim_offsets": true,
|
|
|
|
| 50 |
"extra_special_tokens": {},
|
| 51 |
"mask_token": "<mask>",
|
| 52 |
"model_max_length": 1000000000000000019884624838656,
|
| 53 |
+
"pad_token": "<pad>",
|
| 54 |
"sep_token": "</s>",
|
| 55 |
"tokenizer_class": "BartTokenizer",
|
| 56 |
"trim_offsets": true,
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5496
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:20038f3f3a722c521f78d4841fc191851ead278571cdacedaafa087f11eb7842
|
| 3 |
size 5496
|