mx1 commited on
Commit
032e97b
·
verified ·
1 Parent(s): a21df24

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -11
README.md CHANGED
@@ -24,16 +24,11 @@ Trained on __20T+ tokens of high-quality data__, together with __supervised fine
24
  ### Powerful Complex Reasoning Abilities
25
 
26
  We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
27
-
28
- __Multi-disciplinary knowledge reasoning__: GPQA-Diamond, MMLU-Pro
29
-
30
- __Advanced mathematical reasoning__: AIME 2025, Omni-MATH, OptMATH (advanced mathematical optimization tasks)
31
-
32
- ● __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
33
-
34
- ● __Logical reasoning__: KOR-Bench, ARC-Prize
35
-
36
- ● __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
37
 
38
  Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
39
  <p align="center">
@@ -244,7 +239,7 @@ More usage can be found [here](https://docs.sglang.ai/basic_usage/send_request.h
244
 
245
  ### Finetuning
246
 
247
- We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md). In addition to that, you can also use [Megatron for finetuning](https://github.com/inclusionAI/Ling-V2/blob/main/docs/megatron_sft_training.md).
248
 
249
  ## License
250
 
 
24
  ### Powerful Complex Reasoning Abilities
25
 
26
  We conducted a comprehensive evaluation of Ling-flash-2.0’s reasoning capabilities, reporting strong results on representative benchmarks:
27
+ * __Multi-disciplinary knowledge reasoning__: GPQA-Diamond, MMLU-Pro
28
+ * __Advanced mathematical reasoning__: AIME 2025, Omni-MATH, OptMATH (advanced mathematical optimization tasks)
29
+ * __Challenging code generation__: LiveCodeBench v6, CodeForces-Elo
30
+ * __Logical reasoning__: KOR-Bench, ARC-Prize
31
+ * __Key regulated industries (Finance, Healthcare)__: FinanceReasoning, HealthBench
 
 
 
 
 
32
 
33
  Compared with __dense models under 40B__ (e.g., Qwen3-32B-Non-Thinking, Seed-OSS-36B-Instruct (think budget=0)) and __larger-activation/total-parameter MoE models__ (e.g., Hunyuan-A13B-Instruct, GPT-OSS-120B/low), __Ling-flash-2.0__ demonstrates stronger complex reasoning power. Moreover, it shows high competitiveness on __creative tasks__ (Creative Writing v3).
34
  <p align="center">
 
239
 
240
  ### Finetuning
241
 
242
+ We recommend you to use [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory) to [finetune Ling](https://github.com/inclusionAI/Ling-V2/blob/main/docs/llamafactory_finetuning.md).
243
 
244
  ## License
245