judithspd commited on
Commit
093aa7a
·
verified ·
1 Parent(s): e674e00

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -1
README.md CHANGED
@@ -5,4 +5,54 @@ language:
5
  base_model:
6
  - mistralai/Mistral-7B-Instruct-v0.3
7
  pipeline_tag: question-answering
8
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  base_model:
6
  - mistralai/Mistral-7B-Instruct-v0.3
7
  pipeline_tag: question-answering
8
+ ---
9
+
10
+ # Mistral-7B-Instruct-v0.3-EOSC
11
+
12
+ Federated fine tuned version using data from the EOSC registry.
13
+
14
+ Federated training configuration:
15
+
16
+ - model.name = "mistralai/Mistral-7B-Instruct-v0.3"
17
+ - model.quantization = 4
18
+ - model.gradient-checkpointing = true
19
+ - model.lora.peft-lora-r = 32
20
+ - model.lora.peft-lora-alpha = 64
21
+ - train.save-every-round = 5
22
+ - train.learning-rate-max = 5e-5
23
+ - train.learning-rate-min = 1e-6
24
+ - train.seq-length = 512
25
+ - train.training-arguments.per-device-train-batch-size = 16
26
+ - train.training-arguments.gradient-accumulation-steps = 1
27
+ - train.training-arguments.logging-steps = 10
28
+ - train.training-arguments.num-train-epochs = 2
29
+ - train.training-arguments.max-steps = 10
30
+ - train.training-arguments.save-steps = 1000
31
+ - train.training-arguments.save-total-limit = 10
32
+ - train.training-arguments.gradient-checkpointing = true
33
+ - train.training-arguments.lr-scheduler-type = "constant"
34
+ - strategy.fraction-fit = 0.1
35
+ - strategy.fraction-evaluate = 0.0
36
+ - num-server-rounds = 10
37
+
38
+ The PEFT presented in this model corresponds to 5 rounds of the FL training,
39
+
40
+ The following `bitsandbytes` quantization config was used during training:
41
+ - quant_method: QuantizationMethod.BITS_AND_BYTES
42
+ - _load_in_8bit: False
43
+ - _load_in_4bit: True
44
+ - llm_int8_threshold: 6.0
45
+ - llm_int8_skip_modules: None
46
+ - llm_int8_enable_fp32_cpu_offload: False
47
+ - llm_int8_has_fp16_weight: False
48
+ - bnb_4bit_quant_type: fp4
49
+ - bnb_4bit_use_double_quant: False
50
+ - bnb_4bit_compute_dtype: float32
51
+ - bnb_4bit_quant_storage: uint8
52
+ - load_in_4bit: True
53
+ - load_in_8bit: False
54
+
55
+ ### Framework versions
56
+
57
+
58
+ - PEFT 0.6.2