Update README.md
Browse files
README.md
CHANGED
@@ -24,16 +24,11 @@ Trained on __20T+ tokens of high-quality data__, together with __supervised fine
|
|
24 |
### Powerful Complex Reasoning Abilities
|
25 |
|
26 |
We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
● __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
|
33 |
-
|
34 |
-
● __Logical reasoning__: KOR-Bench, ARC-Prize
|
35 |
-
|
36 |
-
● __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
|
37 |
|
38 |
Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
|
39 |
<p align="center">
|
@@ -244,7 +239,7 @@ More usage can be found [here](https://docs.sglang.ai/basic_usage/send_request.h
|
|
244 |
|
245 |
### Finetuning
|
246 |
|
247 |
-
We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md).
|
248 |
|
249 |
## License
|
250 |
|
|
|
24 |
### Powerful Complex Reasoning Abilities
|
25 |
|
26 |
We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
|
27 |
+
* __Multi-disciplinary knowledge reasoning__: GPQA-Diamond, MMLU-Pro
|
28 |
+
* __Advanced mathematical reasoning__: AIME 2025, Omni-MATH, OptMATH (advanced mathematical optimization tasks)
|
29 |
+
* __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
|
30 |
+
* __Logical reasoning__: KOR-Bench, ARC-Prize
|
31 |
+
* __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
|
34 |
<p align="center">
|
|
|
239 |
|
240 |
### Finetuning
|
241 |
|
242 |
+
We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md).
|
243 |
|
244 |
## License
|
245 |
|