ABSOSUM Phase 2 V1.0 - Weight-Aware Multi-Answer Summarization

This model is a weight-aware T5-based model fine-tuned for multi-answer summarization on Vietnamese Q&A data (ABSOSUM Phase 2).

Model Description

  • Base Model: T5-base
  • Architecture: V2++ with Weight-Aware Cross-Attention
  • Task: Multi-answer summarization with answer importance weighting
  • Language: Vietnamese

Training Details

  • Special tokens: <POST>, <ANS>
  • Max sequence length: 512
  • Max target length: 400
  • Weight injection: log-scaled weights in cross-attention

Performance

ROUGE Scores on Test Set:

  • ROUGE-1: 45.07%
  • ROUGE-2: 21.36%
  • ROUGE-L: 33.30%

Usage

from transformers import T5Tokenizer, T5ForConditionalGeneration

model = T5ForConditionalGeneration.from_pretrained("HuyTran1301/ABSOSUM_Phase2_v1.0")
tokenizer = T5Tokenizer.from_pretrained("HuyTran1301/ABSOSUM_Phase2_v1.0")

# Format input with special tokens
input_text = "<POST> Your question here </s> <ANS> Answer 1 </s> <ANS> Answer 2 </s>"
input_ids = tokenizer(input_text, return_tensors="pt").input_ids

# Generate summary
outputs = model.generate(input_ids, max_length=150)
summary = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(summary)

Citation

If you use this model, please cite:

@misc{absosum_phase2_v1,
  title={ABSOSUM Phase 2: Weight-Aware Multi-Answer Summarization},
  author={Huy Tran},
  year={2025},
  url={https://huggingface.co/HuyTran1301/ABSOSUM_Phase2_v1.0}
}

Training Date

November 28, 2025

Downloads last month
25
Safetensors
Model size
0.2B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support