lisaterumi commited on
Commit
759d1d2
·
verified ·
1 Parent(s): 40560ab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -10
README.md CHANGED
@@ -13,32 +13,26 @@ tags:
13
 
14
  ## Model Description:
15
 
16
- This model is a fine-tuned version of Google's MedGemma, specialized for abstractive summarization of clinical case reports in Portuguese. It was developed as part of our submission to the MultiClinSum 2025 shared task (Portuguese track), organized under the BioASQ Lab at CLEF.
17
 
18
  Despite being compact (4B parameters), the model achieved strong semantic alignment with expert-generated summaries, as measured by BERTScore, and competitive results overall when compared to larger instruction-tuned models in zero-shot settings.
19
 
20
  ## Training Details:
21
 
22
- Base model: https://huggingface.co/unsloth/medgemma-4b-it
23
-
24
- Dataset: Subset of the MultiClinSum Portuguese gold dataset (542 examples for training, 50 for validation)
25
-
26
- Framework: Transformers + PEFT + LoRA (via Unsloth)
27
 
28
  ## Use Cases:
29
 
30
  - Clinical case summarization (Portuguese)
31
-
32
  - Biomedical NLP research
33
-
34
  - Low-resource summarization studies
35
 
36
  ## Limitations:
37
 
38
  - Performance may vary outside of the clinical case report domain
39
-
40
  - Sensitive to prompt design
41
-
42
  - Trained on a small subset.
43
 
44
  ## License
 
13
 
14
  ## Model Description:
15
 
16
+ This model is a fine-tuned version of Google's MedGemma, specialized for abstractive summarization of clinical case reports in Portuguese. It was developed as part of our submission to the [MultiClinSum 2025 shared task](https://temu.bsc.es/multiclinsum) (Portuguese track), organized under the [BioASQ Lab at CLEF](https://www.bioasq.org/).
17
 
18
  Despite being compact (4B parameters), the model achieved strong semantic alignment with expert-generated summaries, as measured by BERTScore, and competitive results overall when compared to larger instruction-tuned models in zero-shot settings.
19
 
20
  ## Training Details:
21
 
22
+ - Base model: https://huggingface.co/unsloth/medgemma-4b-it
23
+ - Dataset: Subset of the MultiClinSum Portuguese gold dataset (542 examples for training, 50 for validation)
24
+ - Framework: Transformers + PEFT + LoRA (via Unsloth)
 
 
25
 
26
  ## Use Cases:
27
 
28
  - Clinical case summarization (Portuguese)
 
29
  - Biomedical NLP research
 
30
  - Low-resource summarization studies
31
 
32
  ## Limitations:
33
 
34
  - Performance may vary outside of the clinical case report domain
 
35
  - Sensitive to prompt design
 
36
  - Trained on a small subset.
37
 
38
  ## License