Update README.md
Browse files
README.md
CHANGED
|
@@ -13,32 +13,26 @@ tags:
|
|
| 13 |
|
| 14 |
## Model Description:
|
| 15 |
|
| 16 |
-
This model is a fine-tuned version of Google's MedGemma, specialized for abstractive summarization of clinical case reports in Portuguese. It was developed as part of our submission to the MultiClinSum 2025 shared task (Portuguese track), organized under the BioASQ Lab at CLEF.
|
| 17 |
|
| 18 |
Despite being compact (4B parameters), the model achieved strong semantic alignment with expert-generated summaries, as measured by BERTScore, and competitive results overall when compared to larger instruction-tuned models in zero-shot settings.
|
| 19 |
|
| 20 |
## Training Details:
|
| 21 |
|
| 22 |
-
Base model: https://huggingface.co/unsloth/medgemma-4b-it
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
Framework: Transformers + PEFT + LoRA (via Unsloth)
|
| 27 |
|
| 28 |
## Use Cases:
|
| 29 |
|
| 30 |
- Clinical case summarization (Portuguese)
|
| 31 |
-
|
| 32 |
- Biomedical NLP research
|
| 33 |
-
|
| 34 |
- Low-resource summarization studies
|
| 35 |
|
| 36 |
## Limitations:
|
| 37 |
|
| 38 |
- Performance may vary outside of the clinical case report domain
|
| 39 |
-
|
| 40 |
- Sensitive to prompt design
|
| 41 |
-
|
| 42 |
- Trained on a small subset.
|
| 43 |
|
| 44 |
## License
|
|
|
|
| 13 |
|
| 14 |
## Model Description:
|
| 15 |
|
| 16 |
+
This model is a fine-tuned version of Google's MedGemma, specialized for abstractive summarization of clinical case reports in Portuguese. It was developed as part of our submission to the [MultiClinSum 2025 shared task](https://temu.bsc.es/multiclinsum) (Portuguese track), organized under the [BioASQ Lab at CLEF](https://www.bioasq.org/).
|
| 17 |
|
| 18 |
Despite being compact (4B parameters), the model achieved strong semantic alignment with expert-generated summaries, as measured by BERTScore, and competitive results overall when compared to larger instruction-tuned models in zero-shot settings.
|
| 19 |
|
| 20 |
## Training Details:
|
| 21 |
|
| 22 |
+
- Base model: https://huggingface.co/unsloth/medgemma-4b-it
|
| 23 |
+
- Dataset: Subset of the MultiClinSum Portuguese gold dataset (542 examples for training, 50 for validation)
|
| 24 |
+
- Framework: Transformers + PEFT + LoRA (via Unsloth)
|
|
|
|
|
|
|
| 25 |
|
| 26 |
## Use Cases:
|
| 27 |
|
| 28 |
- Clinical case summarization (Portuguese)
|
|
|
|
| 29 |
- Biomedical NLP research
|
|
|
|
| 30 |
- Low-resource summarization studies
|
| 31 |
|
| 32 |
## Limitations:
|
| 33 |
|
| 34 |
- Performance may vary outside of the clinical case report domain
|
|
|
|
| 35 |
- Sensitive to prompt design
|
|
|
|
| 36 |
- Trained on a small subset.
|
| 37 |
|
| 38 |
## License
|