multi_task_t5_large / config.json
hazemmessamm's picture
Training in progress, epoch 0
ae92908
raw
history blame contribute delete
326 Bytes
{
"architectures": [
"T5ForMultiTasking"
],
"pooling": "max",
"sequence_classification_decoder_loss_scaling": 1.0,
"sequence_classification_loss_scaling": 1.0,
"t5_checkpoint": "google/flan-t5-large",
"token_classification_loss_scaling": 1.0,
"torch_dtype": "float32",
"transformers_version": "4.33.1"
}