grimjim commited on
Commit
aaa347e
·
verified ·
1 Parent(s): 6d5d334

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -53
README.md CHANGED
@@ -1,53 +1,56 @@
1
- ---
2
- base_model:
3
- - mrfakename/mistral-small-3.1-24b-base-2503-hf
4
- - Doctor-Shotgun/MS3.2-24B-Magnum-Diamond
5
- - PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- pipeline_tag: text-generation
11
- license: apache-2.0
12
- ---
13
- # Modicum-of-Doubt-v1-24B
14
-
15
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
-
17
- The model vision component was excised from all merge contributions.
18
-
19
- Creative text generation outputs seem to trend toward the short side, sometimes to the point of feeling choppy, hence the model name. This model is not the most stellar, but the result is interesting, going against the individual tendency of the two contributing models toward longer outputs.
20
-
21
- Tested sampler settings: temperature 1.0, minP 0.02
22
-
23
- ## Merge Details
24
- ### Merge Method
25
-
26
- This model was merged using the [Task Arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mrfakename/mistral-small-3.1-24b-base-2503-hf](https://huggingface.co/mrfakename/mistral-small-3.1-24b-base-2503-hf) as a base.
27
-
28
- ### Models Merged
29
-
30
- The following models were included in the merge:
31
- * [Doctor-Shotgun/MS3.2-24B-Magnum-Diamond](https://huggingface.co/Doctor-Shotgun/MS3.2-24B-Magnum-Diamond)
32
- * [PocketDoc/Dans-PersonalityEngine-V1.3.0-24b](https://huggingface.co/PocketDoc/Dans-PersonalityEngine-V1.3.0-24b)
33
-
34
- ### Configuration
35
-
36
- The following YAML configuration was used to produce this model:
37
-
38
- ```yaml
39
- base_model: mrfakename/mistral-small-3.1-24b-base-2503-hf
40
- dtype: bfloat16
41
- merge_method: task_arithmetic
42
- parameters:
43
- normalize: true
44
- models:
45
- - model: mrfakename/mistral-small-3.1-24b-base-2503-hf
46
- - model: Doctor-Shotgun/MS3.2-24B-Magnum-Diamond
47
- parameters:
48
- weight: 0.5
49
- - model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
50
- parameters:
51
- weight: 0.5
52
-
53
- ```
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - grimjim/Modicum-of-Doubt-v1-24B
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ - exllamav3
9
+ pipeline_tag: text-generation
10
+ license: apache-2.0
11
+ base_model_relation: quantized
12
+ quanted_by: grimjim
13
+ ---
14
+ # Modicum-of-Doubt-v1-24B-4bpw-h6-exl3
15
+
16
+ This is a quant of merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
17
+
18
+ Exllamav3 was used to create a quant at 4bpw with h6. With 16GB VRAM, it's possible to run 16K context at fp16 with some room to spare.
19
+
20
+ The model vision component was excised from all merge contributions.
21
+
22
+ Creative text generation outputs seem to trend toward the short side, sometimes to the point of feeling choppy, hence the model name. This model is not the most stellar, but the result is interesting, going against the individual tendency of the two contributing models toward longer outputs.
23
+
24
+ Tested sampler settings: temperature 1.0, minP 0.02
25
+
26
+ ## Merge Details
27
+ ### Merge Method
28
+
29
+ This model was merged using the [Task Arithmetic](https://arxiv.org/abs/2212.04089) merge method using [mrfakename/mistral-small-3.1-24b-base-2503-hf](https://huggingface.co/mrfakename/mistral-small-3.1-24b-base-2503-hf) as a base.
30
+
31
+ ### Models Merged
32
+
33
+ The following models were included in the merge:
34
+ * [Doctor-Shotgun/MS3.2-24B-Magnum-Diamond](https://huggingface.co/Doctor-Shotgun/MS3.2-24B-Magnum-Diamond)
35
+ * [PocketDoc/Dans-PersonalityEngine-V1.3.0-24b](https://huggingface.co/PocketDoc/Dans-PersonalityEngine-V1.3.0-24b)
36
+
37
+ ### Configuration
38
+
39
+ The following YAML configuration was used to produce this model:
40
+
41
+ ```yaml
42
+ base_model: mrfakename/mistral-small-3.1-24b-base-2503-hf
43
+ dtype: bfloat16
44
+ merge_method: task_arithmetic
45
+ parameters:
46
+ normalize: true
47
+ models:
48
+ - model: mrfakename/mistral-small-3.1-24b-base-2503-hf
49
+ - model: Doctor-Shotgun/MS3.2-24B-Magnum-Diamond
50
+ parameters:
51
+ weight: 0.5
52
+ - model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
53
+ parameters:
54
+ weight: 0.5
55
+
56
+ ```