deb101 commited on
Commit
73fcd62
·
verified ·
1 Parent(s): b3076ad

Model save

Browse files
README.md CHANGED
@@ -16,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: -299806972635812416.0000
20
- - Ndcg: 0.9561
21
- - Ndcg@25: 0.5733
22
- - Precision@25: 0.4457
23
 
24
  ## Model description
25
 
@@ -51,15 +51,15 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Ndcg | Ndcg@25 | Precision@25 |
55
- |:------------------------:|:------:|:----:|:------------------------:|:------:|:-------:|:------------:|
56
- | -937807332288285952.0000 | 1.0 | 44 | -74029561088243888.0000 | 0.9555 | 0.1932 | 0.0400 |
57
- | -92957537119449904.0000 | 1.9711 | 86 | -299806972635812416.0000 | 0.9561 | 0.5733 | 0.4457 |
58
 
59
 
60
  ### Framework versions
61
 
62
  - Transformers 4.49.0
63
- - Pytorch 2.6.0
64
  - Datasets 3.6.0
65
  - Tokenizers 0.21.1
 
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: -240602318374661568.0000
20
+ - Ndcg: 0.9563
21
+ - Ndcg@25: 0.4422
22
+ - Precision@25: 0.2214
23
 
24
  ## Model description
25
 
 
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Ndcg | Ndcg@25 | Precision@25 |
55
+ |:-------------------------:|:------:|:----:|:------------------------:|:------:|:-------:|:------------:|
56
+ | -2330796205703744512.0000 | 1.0 | 44 | -92728505831320080.0000 | 0.9556 | 0.2014 | 0.0400 |
57
+ | -98429772131152688.0000 | 1.9711 | 86 | -240602318374661568.0000 | 0.9563 | 0.4422 | 0.2214 |
58
 
59
 
60
  ### Framework versions
61
 
62
  - Transformers 4.49.0
63
+ - Pytorch 2.7.1+cu126
64
  - Datasets 3.6.0
65
  - Tokenizers 0.21.1
config.json CHANGED
@@ -15905,6 +15905,7 @@
15905
  "num_attention_heads": 32,
15906
  "num_hidden_layers": 32,
15907
  "num_key_value_heads": 8,
 
15908
  "quantization_config": {
15909
  "_load_in_4bit": true,
15910
  "_load_in_8bit": false,
 
15905
  "num_attention_heads": 32,
15906
  "num_hidden_layers": 32,
15907
  "num_key_value_heads": 8,
15908
+ "num_labels": 7942,
15909
  "quantization_config": {
15910
  "_load_in_4bit": true,
15911
  "_load_in_8bit": false,
eval_loss_plot.png CHANGED
eval_ndcg@25_plot.png CHANGED
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71963d816ce7434c7169aa7fc2606cb22f0fda04a9c8f7a983637f02eca2adc9
3
  size 4323010659
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd5f989c75988e88730d7c9804f1edf50eeb897b1ac5ae913d3e539f87b5beae
3
  size 4323010659
train_loss_plot.png CHANGED
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89e3579564b24f58b16d5bc418fc04f168686996e1bff9fd0a1447ccece4634a
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b213befbacaa13807e9aba070c28ef3a2e97ec3351f16dc389ae217c7e65988
3
+ size 5905