lbourdois commited on
Commit
1316456
·
verified ·
1 Parent(s): 80f56fd

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +66 -53
README.md CHANGED
@@ -1,53 +1,66 @@
1
- ---
2
- base_model:
3
- - arcee-ai/SuperNova-Medius
4
- - arcee-ai/Virtuoso-Small-v2
5
- - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
6
- - Qwen/Qwen2.5-14B-Instruct
7
- - Krystalan/DRT-o1-14B
8
- library_name: transformers
9
- tags:
10
- - mergekit
11
- - merge
12
-
13
- ---
14
- # CoderO1-14B-Preview
15
-
16
- Still experiencing infinite token generation...
17
-
18
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
-
20
- ## Merge Details
21
- ### Merge Method
22
-
23
- This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius) as a base.
24
-
25
- ### Models Merged
26
-
27
- The following models were included in the merge:
28
- * [arcee-ai/Virtuoso-Small-v2](https://huggingface.co/arcee-ai/Virtuoso-Small-v2)
29
- * [deepseek-ai/DeepSeek-R1-Distill-Qwen-14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B)
30
- * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
31
- * [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B)
32
-
33
- ### Configuration
34
-
35
- The following YAML configuration was used to produce this model:
36
-
37
- ```yaml
38
- models:
39
- # Pivot model
40
- - model: arcee-ai/SuperNova-Medius
41
- # Target models
42
- - model: Qwen/Qwen2.5-14B-Instruct
43
- - model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
44
- - model: arcee-ai/Virtuoso-Small-v2
45
- - model: Krystalan/DRT-o1-14B
46
- merge_method: sce
47
- base_model: arcee-ai/SuperNova-Medius
48
- tokenizer_source: base
49
- parameters:
50
- select_topk: 1.0
51
- dtype: bfloat16
52
-
53
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - arcee-ai/SuperNova-Medius
4
+ - arcee-ai/Virtuoso-Small-v2
5
+ - deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
6
+ - Qwen/Qwen2.5-14B-Instruct
7
+ - Krystalan/DRT-o1-14B
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ language:
13
+ - zho
14
+ - eng
15
+ - fra
16
+ - spa
17
+ - por
18
+ - deu
19
+ - ita
20
+ - rus
21
+ - jpn
22
+ - kor
23
+ - vie
24
+ - tha
25
+ - ara
26
+ ---
27
+ # CoderO1-14B-Preview
28
+
29
+ Still experiencing infinite token generation...
30
+
31
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
32
+
33
+ ## Merge Details
34
+ ### Merge Method
35
+
36
+ This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [arcee-ai/Virtuoso-Small-v2](https://huggingface.co/arcee-ai/Virtuoso-Small-v2)
42
+ * [deepseek-ai/DeepSeek-R1-Distill-Qwen-14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B)
43
+ * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
44
+ * [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B)
45
+
46
+ ### Configuration
47
+
48
+ The following YAML configuration was used to produce this model:
49
+
50
+ ```yaml
51
+ models:
52
+ # Pivot model
53
+ - model: arcee-ai/SuperNova-Medius
54
+ # Target models
55
+ - model: Qwen/Qwen2.5-14B-Instruct
56
+ - model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
57
+ - model: arcee-ai/Virtuoso-Small-v2
58
+ - model: Krystalan/DRT-o1-14B
59
+ merge_method: sce
60
+ base_model: arcee-ai/SuperNova-Medius
61
+ tokenizer_source: base
62
+ parameters:
63
+ select_topk: 1.0
64
+ dtype: bfloat16
65
+
66
+ ```