LiangJiang commited on
Commit
aa27fb8
·
verified ·
1 Parent(s): b2c63d0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -17
README.md CHANGED
@@ -88,23 +88,6 @@ To ensure stable and efficient reinforcement learning training for trillion-para
88
 
89
  For the RL training framework, we built a hybrid reward system based on large-scale Serverless Sandbox technology. This system can start up in milliseconds, supports execution environments for over 10 programming languages, and handles request throughput of up to 10K/s. We have open-sourced AReal and hope to accelerate RL training and research in the open-source community through technological openness.
90
 
91
- ## Limitations and Future Plans
92
-
93
- Ring-1T represents the Bailing team’s first attempt at developing a trillion-scale deep reasoning model. The current version may occasionally exhibit issues such as identity recognition bias, language mixing, and repetitive generation. Additionally, since its attention architecture still adopts the GQA approach from Ling 2.0, there remains room for improvement in reasoning efficiency under long-context scenarios.
94
-
95
- We will continue to optimize these aspects in future releases and highly welcome feedback from the community. Furthermore, training for Ring-1T is still ongoing. We are committed to further unlocking the reasoning potential of this trillion-parameter foundation model and look forward to sharing more mature upgraded versions with everyone as soon as possible.
96
-
97
- Welcome to visit our open-source repository and demo page for download and usage.
98
-
99
- Hugging Face: [https://huggingface.co/inclusionAI/Ring-1T](https://huggingface.co/inclusionAI/Ring-1T)
100
-
101
- ModelScope: [https://modelscope.cn/models/inclusionAI/Ring-1T](https://modelscope.cn/models/inclusionAI/Ring-1T)
102
-
103
- Ling Chat (for Chinese users): [https://ling.tbox.cn/chat](https://ling.tbox.cn/chat)
104
-
105
- ZenMux (for overseas developers, offering Chat testing and API capabilities): [https://zenmux.ai/inclusionai/ring-1t](https://zenmux.ai/inclusionai/ring-1t)
106
-
107
- Ring-1T@Aworld IMO test trajectory: [https://github.com/inclusionAI/AWorld/tree/main/examples/imo/samples/samples%20from%20Ring-1T](https://github.com/inclusionAI/AWorld/tree/main/examples/imo/samples/samples%20from%20Ring-1T)
108
 
109
  ## Quickstart
110
 
@@ -194,6 +177,24 @@ More usage can be found [here](https://docs.sglang.ai/basic_usage/send_request.h
194
 
195
  We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ring](https://github.com/inclusionAI/Ring-V2/blob/main/docs/llamafactory_finetuning.md).
196
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
197
  ## License
198
 
199
  This code repository is licensed under [the MIT License](https://github.com/inclusionAI/Ring-V2/blob/master/LICENSE).
 
88
 
89
  For the RL training framework, we built a hybrid reward system based on large-scale Serverless Sandbox technology. This system can start up in milliseconds, supports execution environments for over 10 programming languages, and handles request throughput of up to 10K/s. We have open-sourced AReal and hope to accelerate RL training and research in the open-source community through technological openness.
90
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
 
92
  ## Quickstart
93
 
 
177
 
178
  We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ring](https://github.com/inclusionAI/Ring-V2/blob/main/docs/llamafactory_finetuning.md).
179
 
180
+ ## Limitations and Future Plans
181
+
182
+ Ring-1T represents the Bailing team’s first attempt at developing a trillion-scale deep reasoning model. The current version may occasionally exhibit issues such as identity recognition bias, language mixing, and repetitive generation. Additionally, since its attention architecture still adopts the GQA approach from Ling 2.0, there remains room for improvement in reasoning efficiency under long-context scenarios.
183
+
184
+ We will continue to optimize these aspects in future releases and highly welcome feedback from the community. Furthermore, training for Ring-1T is still ongoing. We are committed to further unlocking the reasoning potential of this trillion-parameter foundation model and look forward to sharing more mature upgraded versions with everyone as soon as possible.
185
+
186
+ Welcome to visit our open-source repository and demo page for download and usage.
187
+
188
+ Hugging Face: [https://huggingface.co/inclusionAI/Ring-1T](https://huggingface.co/inclusionAI/Ring-1T)
189
+
190
+ ModelScope: [https://modelscope.cn/models/inclusionAI/Ring-1T](https://modelscope.cn/models/inclusionAI/Ring-1T)
191
+
192
+ Ling Chat (for Chinese users): [https://ling.tbox.cn/chat](https://ling.tbox.cn/chat)
193
+
194
+ ZenMux (for overseas developers, offering Chat testing and API capabilities): [https://zenmux.ai/inclusionai/ring-1t?utm_source=hf_inclusionAI](https://zenmux.ai/inclusionai/ring-1t?utm_source=hf_inclusionAI)
195
+
196
+ Ring-1T@Aworld IMO test trajectory: [https://github.com/inclusionAI/AWorld/tree/main/examples/imo/samples/samples%20from%20Ring-1T](https://github.com/inclusionAI/AWorld/tree/main/examples/imo/samples/samples%20from%20Ring-1T)
197
+
198
  ## License
199
 
200
  This code repository is licensed under [the MIT License](https://github.com/inclusionAI/Ring-V2/blob/master/LICENSE).