--- license: mit datasets: - saiteja33/DAMASHA language: - en base_model: - FacebookAI/roberta-base - answerdotai/ModernBERT-base pipeline_tag: token-classification --- # DAMASHA-MAS: Mixed-Authorship Adversarial Segmentation (Token Classification) This repository contains a **token-classification model** trained on the **DAMASHA-MAS** benchmark, introduced in: > **DAMASHA: Detecting AI in Mixed Adversarial Texts via Segmentation with Human-interpretable Attribution** The model aims to **segment mixed human–AI text** at *token level* – i.e., decide for each token whether it was written by a *human* or an *LLM*, even under **syntactic adversarial attacks**. - **Base encoders:** - [`FacebookAI/roberta-base`](https://huggingface.co/FacebookAI/roberta-base) - [`answerdotai/ModernBERT-base`](https://huggingface.co/answerdotai/ModernBERT-base) - **Architecture (high level):** RoBERTa + ModernBERT feature fusion → BiGRU + CRF with the **Info-Mask** gating mechanism from the paper. - **Task:** Token classification (binary authorship: human vs AI). - **Language:** English - **License (this model):** MIT - **Training data license:** CC-BY-4.0 via the DAMASHA dataset. If you use this model, **please also cite the DAMASHA paper and dataset** (see Citation section). --- ## 1. Model Highlights - **Fine-grained mixed-authorship detection** Predicts authorship **per token**, allowing reconstruction of human vs AI **spans** in long documents. - **Adversarially robust** Trained and evaluated on **syntactically attacked texts** (misspelling, Unicode substitutions, invisible characters, punctuation swaps, case perturbations, and “all-mixed” attacks). - **Human-interpretable Info-Mask** The architecture incorporates **stylometric features** (perplexity, POS density, punctuation density, lexical diversity, readability) via an **Info-Mask** module that gates token representations in an interpretable way. - **Strong reported performance (from the paper)** On DAMASHA-MAS, the **RMC\*** model (RoBERTa + ModernBERT + CRF + Info-Mask) achieves: - **Token-level**: Accuracy / Precision / Recall / F1 ≈ **0.98** - **Span-level (strict)**: SBDA ≈ **0.45**, SegPre ≈ **0.41** - **Span-level (relaxed IoU ≥ 0.5)**: ≈ **0.82** > ⚠️ The exact numbers for *this* specific checkpoint may differ depending on training run and configuration. The values above are from the paper’s best configuration (RMC\*). --- ## 2. Intended Use ### What this model is for - **Research on human–AI co-authorship** - Studying where LLMs “take over” in mixed texts. - Analysing robustness of detectors under adversarial perturbations. - **Tooling / applications (with human oversight)** - Assisting editors, educators, or moderators to **highlight suspicious spans** rather than making final decisions. - Exploring **interpretability overlays** (e.g., heatmaps over tokens) when combined with Info-Mask outputs. ### What this model is *not* for - Automated “cheating detector” / plagiarism court. - High-stakes decisions affecting people’s livelihood, grades, or reputation **without human review**. - Non-English or heavily code-mixed text (training data is English-centric). Use this model as a **signal**, not a judge. --- ## 3. Data: DAMASHA-MAS The model is trained on the **MAS** benchmark released with the DAMASHA paper and hosted as the Hugging Face dataset: - **Dataset:** [`saiteja33/DAMASHA`](https://huggingface.co/datasets/saiteja33/DAMASHA) ### 3.1 What’s in MAS? MAS consists of **mixed human–AI texts with explicit span tags**: - Human text comes from several corpora for **domain diversity**, including: - Reddit (M4-Reddit) - Yelp & /r/ChangeMyView (MAGE-YELP, MAGE-CMV) - News summaries (XSUM) - Wikipedia (M4-Wiki, MAGE-SQuAD) - ArXiv abstracts (MAGE-SciGen) - QA texts (MAGE-ELI5) - AI text is generated by multiple modern LLMs: - **DeepSeek-V3-671B** (open-source) - **GPT-4o, GPT-4.1, GPT-4.1-mini** (closed-source) ### 3.2 Span tagging Authorship is marked using **explicit tags** around AI spans: - `` … `` denote AI-generated segments within otherwise human text. - The dataset stores text in a `hybrid_text` column, plus metadata such as `has_pair`, and adversarial variants include `attack_name`, `tag_count`, and `attacked_text`. - Tags are sentence-level in annotation, but the model is trained to output **token-level** predictions for finer segmentation. > During training, these tags are converted into **token labels** (2 labels total; see `config.id2label` in the model files). ### 3.3 Adversarial attacks MAS includes multiple **syntactic attacks** applied to the mixed text: - Misspelling - Unicode character substitution - Invisible characters - Punctuation substitution - Upper/lower case swapping - All-mixed combinations of the above These perturbations make tokenization brittle and test robustness of detectors in realistic settings. --- ## 4. Model Architecture & Training ### 4.1 Architecture (conceptual) The model follows the **Info-Mask RMC\*** architecture described in the DAMASHA paper: 1. **Dual encoders** - RoBERTa-base and ModernBERT-base encode the same input sequence. 2. **Feature fusion** - Hidden states from both encoders are fused into a shared representation. 3. **Stylometric Info-Mask** - Hand-crafted style features (perplexity, POS density, punctuation density, lexical diversity, readability) are projected, passed through multi-head attention, and turned into a **scalar mask per token**. - This mask gates the fused encoder states, down-weighting style-irrelevant tokens and emphasizing style-diagnostic ones. :contentReference[oaicite:16]{index=16} 4. **Sequence model + CRF** - A BiGRU layer captures sequential dependencies, followed by a **CRF** layer for structured token labeling with a sequence-level loss. :contentReference[oaicite:17]{index=17} ### 4.2 Training setup (from the paper) Key hyperparameters used for the Info-Mask models on MAS: - **Number of labels:** 2 - **Max sequence length:** 512 - **Batch size:** 64 - **Epochs:** 5 - **Optimizer:** AdamW (with cosine annealing LR schedule) - **Weight decay:** 0.01 - **Gradient clipping:** 1.0 - **Dropout:** Dynamic 0.1–0.3 (initial 0.1) - **Warmup ratio:** 0.1 - **Early stopping patience:** 2 **Hardware & compute** (as reported): - AWS EC2 g6e.xlarge, NVIDIA L40S (48GB) GPU, Ubuntu 24.04 - ≈ 400 GPU hours for experiments. > The exact training script used for this checkpoint is available in the project GitHub: > --- --- license: mit ---