lccurious commited on
Commit
2f9a5a7
·
verified ·
1 Parent(s): 0da4838

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -20,11 +20,10 @@ tags:
20
 
21
  | Benchmark | Ling-mini-2.0 | LLaDA-MoE-7B-A1B-Instruct | LLaDA2.0-mini-preview |
22
  | :------------------------------ | :-------------: | :-------------------------: | :---------------------: |
23
- | **Average** | 64.31 | 53.73 | 60.98 |
24
  | **Knowledge** | | | |
25
  | MMLU | 78.75 | 67.18 | 72.49 |
26
  | MMLU-PRO | 56.40 | 44.64 | 49.22 |
27
- | GPQA | 37.99 | 31.09 | 31.82 |
28
  | CMMLU | 77.84 | 64.30 | 67.53 |
29
  | C-EVAL | 77.85 | 63.93 | 66.54 |
30
  | **Reasoning** | | | |
@@ -36,12 +35,10 @@ tags:
36
  | mbpp | 81.03 | 70.02 | 77.75 |
37
  | MultiPL-E | 62.23 | 52.53 | 62.43 |
38
  | humaneval | 77.44 | 61.59 | 80.49 |
39
- | livecodebench_v6 | 30.18 | 13.27 | 19.93 |
40
  | Bigcodebench-Full | 35.88 | 20.44 | 30.44 |
41
  | **Math** | | | |
42
  | GSM8K | 91.58 | 82.41 | 89.01 |
43
  | math | 82.22 | 58.68 | 73.50 |
44
- | OlympiadBench | 49.93 | 21.04 | 36.67 |
45
  | **Agent & Alignment** | | | |
46
  | BFCL_Live | 45.74 | 63.09 | 74.11 |
47
  | IFEval-strict -prompt | 69.13 | 59.33 | 62.50 |
@@ -60,12 +57,18 @@ Supports **tool calling** and achieves excellent performance in complex agent-ba
60
  + **Open & Extensible**:
61
  Fully open-source with commitment to transparency. We plan to release a **leading inference framework** in the future and continue investing in cutting-edge areas like **diffusion LLMs (dLLM)** to drive disruptive innovation.
62
 
 
 
 
 
 
63
  ---
64
 
65
  ## 📦 Model Variants
66
  | Model ID | Description | Hugging Face Link |
67
  | --- | --- | --- |
68
  | `inclusionAI/LLaDA2.0-mini-preview` | Instruction-tuned model, ready for downstream applications. | [🤗 Model Card](https://huggingface.co/inclusionAI/LLaDA2.0-mini-preview) |
 
69
 
70
 
71
  ---
 
20
 
21
  | Benchmark | Ling-mini-2.0 | LLaDA-MoE-7B-A1B-Instruct | LLaDA2.0-mini-preview |
22
  | :------------------------------ | :-------------: | :-------------------------: | :---------------------: |
23
+ | **Average** | 68.98 | 59.72 | 66.89 |
24
  | **Knowledge** | | | |
25
  | MMLU | 78.75 | 67.18 | 72.49 |
26
  | MMLU-PRO | 56.40 | 44.64 | 49.22 |
 
27
  | CMMLU | 77.84 | 64.30 | 67.53 |
28
  | C-EVAL | 77.85 | 63.93 | 66.54 |
29
  | **Reasoning** | | | |
 
35
  | mbpp | 81.03 | 70.02 | 77.75 |
36
  | MultiPL-E | 62.23 | 52.53 | 62.43 |
37
  | humaneval | 77.44 | 61.59 | 80.49 |
 
38
  | Bigcodebench-Full | 35.88 | 20.44 | 30.44 |
39
  | **Math** | | | |
40
  | GSM8K | 91.58 | 82.41 | 89.01 |
41
  | math | 82.22 | 58.68 | 73.50 |
 
42
  | **Agent & Alignment** | | | |
43
  | BFCL_Live | 45.74 | 63.09 | 74.11 |
44
  | IFEval-strict -prompt | 69.13 | 59.33 | 62.50 |
 
57
  + **Open & Extensible**:
58
  Fully open-source with commitment to transparency. We plan to release a **leading inference framework** in the future and continue investing in cutting-edge areas like **diffusion LLMs (dLLM)** to drive disruptive innovation.
59
 
60
+ ## 🗺️ What's Next
61
+
62
+ + **Supercharged Reasoning with LLaDA 2.0:** LLaDA 2.0 series will be fine-tuned with **Reinforcement Learning**, unlocking a new level of sophisticated reasoning and problem-solving abilities.
63
+ + **Tools for Innovators:** we will release a **detailed tutorial** and our complete **post-training framework**. Whether you want to master the current model or build your own customized versions, you'll have the tools you need. Stay tuned
64
+
65
  ---
66
 
67
  ## 📦 Model Variants
68
  | Model ID | Description | Hugging Face Link |
69
  | --- | --- | --- |
70
  | `inclusionAI/LLaDA2.0-mini-preview` | Instruction-tuned model, ready for downstream applications. | [🤗 Model Card](https://huggingface.co/inclusionAI/LLaDA2.0-mini-preview) |
71
+ | `inclusionAI/LLaDA2.0-flash-preview` | Instruction-tuned model, ready for downstream applications. | [🤗 Model Card](https://huggingface.co/inclusionAI/LLaDA2.0-flash-preview) |
72
 
73
 
74
  ---