Lingaaaaaaa commited on
Commit
6992822
·
verified ·
1 Parent(s): 5d24bf8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -11
README.md CHANGED
@@ -2,6 +2,17 @@
2
  license: mit
3
  ---
4
 
 
 
 
 
 
 
 
 
 
 
 
5
  <p align="center">
6
  <img src="https://github.com/yinjjiew/Data/raw/main/dllm-rl/figure1.png" width="100%"/>
7
  </p>
@@ -12,22 +23,15 @@ license: mit
12
  </p>
13
 
14
 
15
- # Introduction to TraDo
16
-
17
- We introduce **TraDo**, SOTA diffusion language model, trained with **TraceRL**.
18
-
19
- * **TraDo-4B-Instruct** and **TraDo-8B-Instruct** outperform similarly sized strong AR models across math reasoning tasks.
20
- * **TraDo-8B-Thinking** is the first Long-CoT diffusion language model.
21
 
22
- [Paper](https://arxiv.org/abs/2506.03136) | [Code](https://github.com/Gen-Verse/dLLM-RL)
23
 
24
  # Citation
25
 
26
  ```
27
- @article{wang2025cure,
28
- title={Co-Evolving LLM Coder and Unit Tester via Reinforcement Learning},
29
- author={Wang, Yinjie and Yang, Ling and Tian, Ye and Shen, Ke and Wang, Mengdi},
30
- journal={arXiv preprint arXiv:2506.03136},
31
  year={2025}
32
  }
33
  ```
 
2
  license: mit
3
  ---
4
 
5
+ # Introduction to TraDo
6
+
7
+ [Paper](https://arxiv.org/abs/2509.06949) | [Code](https://github.com/Gen-Verse/dLLM-RL)
8
+
9
+ We introduce **TraDo**, SOTA diffusion language model, trained with **TraceRL**.
10
+
11
+ * **TraDo-4B-Instruct** and **TraDo-8B-Instruct** outperform similarly sized strong AR models across math reasoning tasks.
12
+ * **TraDo-8B-Thinking** is the first Long-CoT diffusion language model.
13
+
14
+
15
+
16
  <p align="center">
17
  <img src="https://github.com/yinjjiew/Data/raw/main/dllm-rl/figure1.png" width="100%"/>
18
  </p>
 
23
  </p>
24
 
25
 
 
 
 
 
 
 
26
 
 
27
 
28
  # Citation
29
 
30
  ```
31
+ @article{wang2025trado,
32
+ title={Revolutionizing Reinforcement Learning Framework for Diffusion Large Language Models},
33
+ author={Wang, Yinjie and Yang, Ling and Li, Bowen and Tian, Ye and Shen, Ke and Wang, Mengdi},
34
+ journal={arXiv preprint arXiv:2509.06949},
35
  year={2025}
36
  }
37
  ```